Search results for: probability bivariant models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7585

Search results for: probability bivariant models

4195 Computation of Drag and Lift Coefficients on Submerged Vanes in Open Channels

Authors: Anshul Jain, P. Deepak Kumar, P. K. S. Dikshit

Abstract:

To stabilize the riverbanks in the curved reaches of alluvial channels due to erosion and to stop sediment transportation, many models and theories have been put forth. One among such methods is to install flat vanes on the channel bed in predetermined manner. In practical, a relatively small no of vanes can produce bend flows which are practically uniform across the channel. The objective of the present study is to measure the drag and lift on such submerged vanes in open channels. Experiments were performed and the data collected have been presented and analyzed. Using the data collected herein, predictors for the coefficients of drag and lift have been developed. Such predictors yield the value of these coefficients for the known fluid properties and flow characteristic of the channel.

Keywords: drag, lift, vanes, open channel

Procedia PDF Downloads 337
4194 An Examination of Earnings Management by Publicly Listed Targets Ahead of Mergers and Acquisitions

Authors: T. Elrazaz

Abstract:

This paper examines accrual and real earnings management by publicly listed targets around mergers and acquisitions. Prior literature shows that earnings management around mergers and acquisitions can have a significant economic impact because of the associated wealth transfers among stakeholders. More importantly, acting on behalf of their shareholders or pursuing their self-interests, managers of both targets and acquirers may be equally motivated to manipulate earnings prior to an acquisition to generate higher gains for their shareholders or themselves. Building on the grounds of information asymmetry, agency conflicts, stewardship theory, and the revelation principle, this study addresses the question of whether takeover targets employ accrual and real earnings management in the periods prior to the announcement of Mergers and Acquisitions (M&A). Additionally, this study examines whether acquirers are able to detect targets’ earnings management, and in response, adjust the acquisition premium paid in order not to face the risk of overpayment. This study uses an aggregate accruals approach in estimating accrual earnings management as proxied by estimated abnormal accruals. Additionally, real earnings management is proxied for by employing widely used models in accounting and finance literature. The results of this study indicate that takeover targets manipulate their earnings using accruals in the second year with an earnings release prior to the announcement of the M&A. Moreover, in partitioning the sample of targets according to the method of payment used in the deal, the results are restricted only to targets of stock-financed deals. These results are consistent with the argument that targets of cash-only or mixed-payment deals do not have the same strong motivations to manage their earnings as their stock-financed deals counterparts do additionally supporting the findings of prior studies that the method of payment in takeovers is value relevant. The findings of this study also indicate that takeover targets manipulate earnings upwards through cutting discretionary expenses the year prior to the acquisition while they do not do so by manipulating sales or production costs. Moreover, in partitioning the sample of targets according to the method of payment used in the deal, the results are restricted only to targets of stock-financed deals, providing further robustness to the results derived under the accrual-based models. Finally, this study finds evidence suggesting that acquirers are fully aware of the accrual-based techniques employed by takeover targets and can unveil such manipulation practices. These results are robust to alternative accrual and real earnings management proxies, as well as controlling for the method of payment in the deal.

Keywords: accrual earnings management, acquisition premium, real earnings management, takeover targets

Procedia PDF Downloads 101
4193 Optimization and Evaluation of 177lu-Dotatoc as a Potential Agent for Peptide Receptor Radionuclide Therapy

Authors: H. Yousefnia, MS. Mousavi-Daramoroudi, S. Zolghadri, F. Abbasi-Davani

Abstract:

High expression of somatostatin receptors on a wide range of human tumours makes them as potential targets for peptide receptor radionuclide tomography. A series of octreotide analogues were synthesized while [DOTA-DPhe1, Tyr3]octreotide (DOTATOC) indicated advantageous properties in tumour models. In this study, 177Lu-DOTATOC was prepared with the radiochemical purity of higher than 99% in 30 min at the optimized condition. Biological behavior of the complex was studied after intravenous injection into the Syrian rats. Major difference uptake was observed compared to 177LuCl3 solution especially in somatostatin receptor-positive tissues such as pancreas and adrenal.

Keywords: Biodistribution, 177Lu, Octreotide, Syrian rats

Procedia PDF Downloads 433
4192 Paraplegic Dimensions of Asymmetric Warfare: A Strategic Analysis for Resilience Policy Plan

Authors: Sehrish Qayyum

Abstract:

In this age of constant technology, asymmetrical warfare could not be won. Attuned psychometric study confirms that screaming sometimes is more productive than active retaliation against strong adversaries. Asymmetric warfare is a game of nerves and thoughts with least vigorous participation for large anticipated losses. It creates the condition of paraplegia with partial but permanent immobility, which effects the core warfare operations, being screams rather than active retaliation. When one’s own power is doubted, it gives power to one’s own doubt to ruin all planning either done with superlative cost-benefit analysis. Strategically calculated estimation of asymmetric warfare since the early WWI to WWII, WWII-to Cold War, and then to the current era in three chronological periods exposits that courage makes nations win the battle of warriors to battle of comrades. Asymmetric warfare has been most difficult to fight and survive due to unexpectedness and being lethal despite preparations. Thoughts before action may be the best-assumed strategy to mix Regional Security Complex Theory and OODA loop to develop the Paraplegic Resilience Policy Plan (PRPP) to win asymmetric warfare. PRPP may serve to control and halt the ongoing wave of terrorism, guerilla warfare, and insurgencies, etc. PRPP, along with a strategic work plan, is based on psychometric analysis to deal with any possible war condition and tactic to save millions of innocent lives such that lost in Christchurch New Zealand in 2019, November 2015 Paris attacks, and Berlin market attacks in 2016, etc. Getting tangled into self-imposed epistemic dilemmas results in regret that becomes the only option of performance. It is a descriptive psychometric analysis of war conditions with generic application of probability tests to find the best possible options and conditions to develop PRPP for any adverse condition possible so far. Innovation in technology begets innovation in planning and action-plan to serve as a rheostat approach to deal with asymmetric warfare.

Keywords: asymmetric warfare, psychometric analysis, PRPP, security

Procedia PDF Downloads 123
4191 Estimating the Technological Deviation Impact on the Value of the Output Parameter of the Induction Converter

Authors: Marinka K. Baghdasaryan, Siranush M. Muradyan, Avgen A. Gasparyan

Abstract:

Based on the experimental data, the impact of resistance and reactance of the winding, as well as the magnetic permeability of the magnetic circuit steel material on the value of the electromotive force of the induction converter is investigated. The obtained results allow to estimate the main technological spreads and determine the maximum level of the electromotive force change. By the method of experiment planning, the expression of a polynomial for the electromotive force which can be used to estimate the adequacy of mathematical models to be used at the investigation and design of induction converters is obtained.

Keywords: induction converter, electromotive force, expectation, technological spread, deviation, planning an experiment, polynomial, confidence level

Procedia PDF Downloads 454
4190 Doing Bad for a Greater Good: Moral Disengagement in Social and Commercial Entrepreneurial Contexts

Authors: Thorsten Auer, Sumaya Islam, Sabrina Plaß, Colin Wooldridge

Abstract:

Whether individuals are more likely to forgo some ethical values if it is for a “great” social mission remains questionable. Research interest in the mechanism of moral disengagement has risen sharply in the organizational context over the last decades. Moral disengagement provides an explanatory approach to why individuals decide against their moral intent and describes the tendency to make unethical decisions due to a lack of self-regulation given various actions and their consequences. In our study, we examine the differences between individual decision-making given a commercial and social entrepreneurial context. Thereby, we investigate whether individuals in a social entrepreneurial context, characterized by pro-social goals and purpose beyond profit maximization, tend to make more or less “unethical” decisions in trade-off situations than those given a profit-focused commercial, entrepreneurial context. While a general priming effect may explain the tendency for individuals to make less unethical decisions given a social context, it remains unclear how individuals decide given a trade-off in that specific context. The trade-off in our study is characterized by the option to decide (un-) ethically to enhance the business purpose (in the social context, a social purpose, in the commercial context, a profit-maximization purpose). To investigate which characteristics of the context –and specifically of a trade-off – lead individuals to disregard and override their ethical values for a “greater good”, we design a conjoint analysis. This approach allows us to vary the attributes and scenarios and to test which attributes of a trade-off increase the probability of making an unethical choice. We add survey data to examine the individual propensity to morally disengage as an influencing factor to prefer certain attributes. Currently, we are in the final process of designing the conjoint analysis and plan to conduct the study by December 2022. We contribute to a better understanding of the role of moral disengagement in individual decision-making in a (social) entrepreneurial trade-off.

Keywords: moral disengagement, social entrepreneurship, unethical decision, conjoint analysis

Procedia PDF Downloads 76
4189 Review and Comparison of Associative Classification Data Mining Approaches

Authors: Suzan Wedyan

Abstract:

Data mining is one of the main phases in the Knowledge Discovery Database (KDD) which is responsible of finding hidden and useful knowledge from databases. There are many different tasks for data mining including regression, pattern recognition, clustering, classification, and association rule. In recent years a promising data mining approach called associative classification (AC) has been proposed, AC integrates classification and association rule discovery to build classification models (classifiers). This paper surveys and critically compares several AC algorithms with reference of the different procedures are used in each algorithm, such as rule learning, rule sorting, rule pruning, classifier building, and class allocation for test cases.

Keywords: associative classification, classification, data mining, learning, rule ranking, rule pruning, prediction

Procedia PDF Downloads 524
4188 Hydrological, Hydraulics, Analysis and Design of the Aposto –Yirgalem Road Upgrading Project, Ethiopia

Authors: Azazhu Wassie

Abstract:

This study tried to analyze and identify the drainage pattern and catchment characteristics of the river basin and assess the impact of the hydrologic parameters (catchment area, rainfall intensity, runoff coefficient, land use, and soil type) on the referenced study area. Since there is no river gauging station near the road, even for large rivers, rainfall-runoff models are adopted for flood estimation, i.e., for catchment areas less than 50 ha, the rational method is used; for catchment areas, less than 65 km², the SCS unit hydrograph method is used; and for catchment areas greater than 65 km², HEC-HMS is adopted for flood estimation.

Keywords: Arc GIS, catchment area, land use/land cover, peak flood, rainfall intensity

Procedia PDF Downloads 18
4187 A Bayesian Approach for Analyzing Academic Article Structure

Authors: Jia-Lien Hsu, Chiung-Wen Chang

Abstract:

Research articles may follow a simple and succinct structure of organizational patterns, called move. For example, considering extended abstracts, we observe that an extended abstract usually consists of five moves, including Background, Aim, Method, Results, and Conclusion. As another example, when publishing articles in PubMed, authors are encouraged to provide a structured abstract, which is an abstract with distinct and labeled sections (e.g., Introduction, Methods, Results, Discussions) for rapid comprehension. This paper introduces a method for computational analysis of move structures (i.e., Background-Purpose-Method-Result-Conclusion) in abstracts and introductions of research documents, instead of manually time-consuming and labor-intensive analysis process. In our approach, sentences in a given abstract and introduction are automatically analyzed and labeled with a specific move (i.e., B-P-M-R-C in this paper) to reveal various rhetorical status. As a result, it is expected that the automatic analytical tool for move structures will facilitate non-native speakers or novice writers to be aware of appropriate move structures and internalize relevant knowledge to improve their writing. In this paper, we propose a Bayesian approach to determine move tags for research articles. The approach consists of two phases, training phase and testing phase. In the training phase, we build a Bayesian model based on a couple of given initial patterns and the corpus, a subset of CiteSeerX. In the beginning, the priori probability of Bayesian model solely relies on initial patterns. Subsequently, with respect to the corpus, we process each document one by one: extract features, determine tags, and update the Bayesian model iteratively. In the testing phase, we compare our results with tags which are manually assigned by the experts. In our experiments, the promising accuracy of the proposed approach reaches 56%.

Keywords: academic English writing, assisted writing, move tag analysis, Bayesian approach

Procedia PDF Downloads 319
4186 A Contemporary Advertising Strategy on Social Networking Sites

Authors: M. S. Aparna, Pushparaj Shetty D.

Abstract:

Nowadays social networking sites have become so popular that the producers or the sellers look for these sites as one of the best options to target the right audience to market their products. There are several tools available to monitor or analyze the social networks. Our task is to identify the right community web pages and find out the behavior analysis of the members by using these tools and formulate an appropriate strategy to market the products or services to achieve the set goals. The advertising becomes more effective when the information of the product/ services come from a known source. The strategy explores great buying influence in the audience on referral marketing. Our methodology proceeds with critical budget analysis and promotes viral influence propagation. In this context, we encompass the vital bits of budget evaluation such as the number of optimal seed nodes or primary influential users activated onset, an estimate coverage spread of nodes and maximum influence propagating distance from an initial seed to an end node. Our proposal for Buyer Prediction mathematical model arises from the urge to perform complex analysis when the probability density estimates of reliable factors are not known or difficult to calculate. Order Statistics and Buyer Prediction mapping function guarantee the selection of optimal influential users at each level. We exercise an efficient tactics of practicing community pages and user behavior to determine the product enthusiasts on social networks. Our approach is promising and should be an elementary choice when there is little or no prior knowledge on the distribution of potential buyers on social networks. In this strategy, product news propagates to influential users on or surrounding networks. By applying the same technique, a user can search friends who are capable to advise better or give referrals, if a product interests him.

Keywords: viral marketing, social network analysis, community web pages, buyer prediction, influence propagation, budget constraints

Procedia PDF Downloads 247
4185 Comparative Study of Mutations Associated with Second Line Drug Resistance and Genetic Background of Mycobacterium tuberculosis Strains

Authors: Syed Beenish Rufai, Sarman Singh

Abstract:

Background: Performance of Genotype MTBDRsl (Hain Life science GmbH Germany) for detection of mutations associated with second-line drug resistance is well known. However, less evidence regarding the association of mutations and genetic background of strains is known which, in the future, is essential for clinical management of anti-tuberculosis drugs in those settings where the probability of particular genotype is predominant. Material and Methods: During this retrospective study, a total of 259 MDR-TB isolates obtained from pulmonary TB patients were tested for second-line drug susceptibility testing (DST) using Genotype MTBDRsl VER 1.0 and compared with BACTEC MGIT-960 as a reference standard. All isolates were further characterized using spoligotyping. The spoligo patterns obtained were compared and analyzed using SITVIT_WEB. Results: Of total 259 MDR-TB isolates which were screened for second-line DST by Genotype MTBDRsl, mutations were found to be associated with gyrA, rrs and emb genes in 82 (31.6%), 2 (0.8%) and 90 (34.7%) isolates respectively. 16 (6.1%) isolates detected mutations associated with both FQ as well as to AG/CP drugs (XDR-TB). No mutations were detected in 159 (61.4%) isolates for corresponding gyrA and rrs genes. Genotype MTBDRsl showed a concordance of 96.4% for detection of sensitive isolates in comparison with second-line DST by BACTEC MGIT-960 and 94.1%, 93.5%, 60.5% and 50% for detection of XDR-TB, FQ, EMB, and AMK/CAP respectively. D94G was the most prevalent mutation found among (38 (46.4%)) OFXR isolates (37 FQ mono-resistant and 1 XDR-TB) followed by A90V (23 (28.1%)) (17 FQ mono-resistant and 6 XDR-TB). Among AG/CP resistant isolates A1401G was the most frequent mutation observed among (11 (61.1%)) isolates (2 AG/CP mono-resistant isolates and 9 XDR-TB isolates) followed by WT+A1401G (6 (33.3%)) and G1484T (1 (5.5%)) respectively. On spoligotyping analysis, Beijing strain (46%) was found to be the most predominant strain among pre-XDR and XDR TB isolates followed by CAS (30%), X (6%), Unique (5%), EAI and T each of 4%, Manu (3%) and Ural (2%) respectively. Beijing strain was found to be strongly associated with D94G (47.3%) and A90V mutations by (47.3%) and 34.8% followed by CAS strain by (31.6%) and 30.4% respectively. However, among AG/CP resistant isolates, only Beijing strain was found to be strongly associated with A1401G and WT+A1401G mutations by 54.5% and 50% respectively. Conclusion: Beijing strain was found to be strongly associated with the most prevalent mutations among pre-XDR and XDR TB isolates. Acknowledgments: Study was supported with Grant by All India Institute of Medical Sciences, New Delhi reference No. P-2012/12452.

Keywords: tuberculosis, line probe assay, XDR TB, drug susceptibility

Procedia PDF Downloads 128
4184 Robust Variogram Fitting Using Non-Linear Rank-Based Estimators

Authors: Hazem M. Al-Mofleh, John E. Daniels, Joseph W. McKean

Abstract:

In this paper numerous robust fitting procedures are considered in estimating spatial variograms. In spatial statistics, the conventional variogram fitting procedure (non-linear weighted least squares) suffers from the same outlier problem that has plagued this method from its inception. Even a 3-parameter model, like the variogram, can be adversely affected by a single outlier. This paper uses the Hogg-Type adaptive procedures to select an optimal score function for a rank-based estimator for these non-linear models. Numeric examples and simulation studies will demonstrate the robustness, utility, efficiency, and validity of these estimates.

Keywords: asymptotic relative efficiency, non-linear rank-based, rank estimates, variogram

Procedia PDF Downloads 413
4183 Wireless Battery Charger with Adaptive Rapid-Charging Algorithm

Authors: Byoung-Hee Lee

Abstract:

Wireless battery charger with adaptive rapid charging algorithm is proposed. The proposed wireless charger adopts voltage regulation technique to reduce the number of power conversion steps. Moreover, based on battery models, an adaptive rapid charging algorithm for Li-ion batteries is obtained. Rapid-charging performance with the proposed wireless battery charger and the proposed rapid charging algorithm has been experimentally verified to show more than 70% charging time reduction compared to conventional constant-current constant-voltage (CC-CV) charging without the degradation of battery lifetime.

Keywords: wireless, battery charger, adaptive, rapid-charging

Procedia PDF Downloads 356
4182 Highly Active, Non-Platinum Metal Catalyst Material as Bi-Functional Air Cathode in Zinc Air Battery

Authors: Thirupathi Thippani, Kothandaraman Ramanujam

Abstract:

Current research on energy storage has been paid to metal-air batteries, because of attractive alternate energy source for the future. Metal – air batteries have the probability to significantly increase the power density, decrease the cost of energy storage and also used for a long time due to its high energy density, low-level pollution, light weight. The performance of these batteries mostly restricted by the slow kinetics of the oxygen reduction reaction (ORR) and oxygen evolution reaction (OER) on cathode during battery discharge and charge. The ORR and OER are conventionally carried out with precious metals (such as Pt) and metal oxides (such as RuO₂ and IrO₂) as catalysts separately. However, these metal-based catalysts are regularly undergoing some difficulties, including high cost, low selectivity, poor stability and unfavorable to environmental effects. So, in order to develop the active, stable, corrosion resistance and inexpensive bi-functional catalyst material is mandatory for the commercialization of zinc-air rechargeable battery technology. We have attempted and synthesized non-precious metal (NPM) catalysts comprising cobalt and N-doped multiwalled carbon nanotubes (N-MWCNTs-Co) were synthesized by the solid-state pyrolysis (SSP) of melamine with Co₃O₄. N-MWCNTs-Co acts as an excellent electrocatalyst for both the oxygen reduction reaction (ORR) and the oxygen evolution reaction (OER), and hence can be used in secondary metal-air batteries and in unitized regenerative fuel cells. It is important to study the OER and ORR at high concentrations of KOH as most of the metal-air batteries employ KOH concentrations > 4M. In the first 16 cycles of the zinc-air battery while using N-MWCNTs-Co, 20 wt.% Pt/C or 20 wt.% IrO₂/C as air electrodes. In the ORR regime (the discharge profile of the zinc-air battery), the cell voltage exhibited by N-MWCNTs-Co was 44 and 83 mV higher (based on 5th cycle) in comparison to of 20 wt.% Pt/C and 20 wt.% IrO₂/C respectively. To demonstrate this promise, a zinc-air battery was assembled and tested at a current density of 0.5 Ag⁻¹ for charge-discharge 100 cycles.

Keywords: oxygen reduction reaction (ORR), oxygen evolution reaction(OER), non-platinum, zinc air battery

Procedia PDF Downloads 222
4181 Thermodynamic Behaviour of Binary Mixtures of 1, 2-Dichloroethane with Some Cyclic Ethers: Experimental Results and Modelling

Authors: Fouzia Amireche-Ziar, Ilham Mokbel, Jacques Jose

Abstract:

The vapour pressures of the three binary mixtures: 1, 2- dichloroethane + 1,3-dioxolane, + 1,4-dioxane or + tetrahydropyrane, are carried out at ten temperatures ranging from 273 to 353.15 K. An accurate static device was employed for these measurements. The VLE data were reduced using the Redlich-Kister equation by taking into consideration the vapour pressure non-ideality in terms of the second molar virial coefficient. The experimental data were compared to the results predicted with the DISQUAC and Dortmund UNIFAC group contribution models for the total pressures P and the excess molar Gibbs energies GE.

Keywords: disquac model, dortmund UNIFAC model, excess molar Gibbs energies GE, VLE

Procedia PDF Downloads 250
4180 Modelling Structural Breaks in Stock Price Time Series Using Stochastic Differential Equations

Authors: Daniil Karzanov

Abstract:

This paper studies the effect of quarterly earnings reports on the stock price. The profitability of the stock is modeled by geometric Brownian diffusion and the Constant Elasticity of Variance model. We fit several variations of stochastic differential equations to the pre-and after-report period using the Maximum Likelihood Estimation and Grid Search of parameters method. By examining the change in the model parameters after reports’ publication, the study reveals that the reports have enough evidence to be a structural breakpoint, meaning that all the forecast models exploited are not applicable for forecasting and should be refitted shortly.

Keywords: stock market, earnings reports, financial time series, structural breaks, stochastic differential equations

Procedia PDF Downloads 190
4179 The Role of Technology in Transforming the Finance, Banking, and Insurance Sectors

Authors: Farid Fahami

Abstract:

This article explores the transformative role of technology in the finance, banking, and insurance sectors. It examines key technological trends such as AI, blockchain, data analytics, and digital platforms and their impact on operations, customer experiences, and business models. The article highlights the benefits of technology adoption, including improved efficiency, cost reduction, enhanced customer experiences, and expanded financial inclusion. It also addresses challenges like cybersecurity, data privacy, and the need for upskilling. Real-world case studies demonstrate successful technology integration, and recommendations for stakeholders emphasize embracing innovation and collaboration. The article concludes by emphasizing the importance of technology in shaping the future of these sectors.

Keywords: banking, finance, insurance, technology

Procedia PDF Downloads 61
4178 Adolescent Sleep Hygiene Scale and Adolescent Sleep Wake Scale: Factorial Analysis and Validation for Indian Population

Authors: Sataroopa Mishra, Mona Basker, Sneha Varkki, Ram Kumar Pandian, Grace Rebekah

Abstract:

Background: Sleep deprivation is a matter of public health importance among adolescents. We used adolescent sleep wake scale and adolescent sleep hygiene scale to determine the sleep quality and sleep hygiene respectively of school going adolescents in Vellore city of India. The objective of the study was to do factorial analysis of the scales and validate it for use in local population. Methods: Observational questionnaire based cross sectional study. Setting: Community based school survey in a semi-urban setting in three schools in Vellore city. Data collection: Non probability sample was collected form students studying in standard 9 and 11. Students filled Adolescent Sleep Wake scale (ASWS) and Adolescent Sleep Hygiene Scale (ASHS) translated into vernacular language. Data Analysis: Exploratory Factorial Analysis was used to see the factor loading of various components of the two scales. Confirmatory factorial analysis is subsequently planned for assessing the internal validity of the scales.Results: 557 adolescents were included in the study of 12 – 17 years old. Exploratory factorial analysis of adolescent sleep hygiene scale indicated significant factor loading for 18 items from 28 items originally devised by the authors and has been reconstructed to four domains instead of 9 domains in the original scale namely sleep stability, cognitive – emotional, Physiological - bed time routine - behavioural arousal factor (activites before bedtime and during bed time), Sleep environment (lighting and bed sharing). Factorial analysis of Adolescent sleep wake scale showed factor loading of 18 items out of 28 items in original scale reconstructed into 5 aspects of sleep quality. Conclusions: The factorial analysis gives a reconstructed scale useful for the local population. Further a confirmatory factorial analysis has been subsequently planned to determine the internal consistency of the scale for local population.

Keywords: factorial analysis, sleep hygiene, sleep quality, adolescent sleep scale

Procedia PDF Downloads 268
4177 Entrepreneurial Leadership in a Startup Context: A Comparative Study on Two Egyptian Startup Businesses

Authors: Nada Basset

Abstract:

Problem Statement: The study examines the important role of leading change inside start-ups and highlights the challenges faced by an entrepreneur during the startup phase of the business. Research Methods/Procedures/Approaches: A qualitative research approach is taken, using the case study analysis method. A comparative study was made between two day care nurseries in Greater Cairo. Non-probability purposive sampling was used and a triangulation of semi-structured interviews, document analysis and participant-observation were applied simultaneously. The in-depth case study analysis took place over a longitudinal study of four calendar months. Results/Findings: Findings demonstrated that leading change in an entrepreneurial setup must be initiated by the entrepreneur, who must also be the owner of the change process. Another important finding showed that the culture of change, although created by the entrepreneur, needs the support and engagement of followers, who should be sharing the same value system and vision of the entrepreneur. Conclusions and Implications: An important implication suggests that during the first year of a start-up lifecycle, special emphasis must be made to the recruitment and selection of personnel, who should play a role into setting the new start-up culture and help it grow or shrink. Another drawn conclusion is that the success of the change must be measured in both quantitative and qualitative terms. Increasing revenues and customer attrition rates -as quantitative KPIs- must be aligned with other qualitative KPIs like customer satisfaction, employee satisfaction, and organizational commitment and business reputation. Originality of Paper: The paper addresses change management in an entrepreneurial concept, with an empirical application on an Egyptian start-up model providing a service to both adults and children. This privileges the research as the constructs measured merged together the level of satisfaction of employees, decision-makers (parents of children), and the users (children).

Keywords: leadership, change management, entrepreneurship, startup business

Procedia PDF Downloads 168
4176 Calibration of 2D and 3D Optical Measuring Instruments in Industrial Environments at Submillimeter Range

Authors: Alberto Mínguez-Martínez, Jesús de Vicente y Oliva

Abstract:

Modern manufacturing processes have led to the miniaturization of systems and, as a result, parts at the micro-and nanoscale are produced. This trend seems to become increasingly important in the near future. Besides, as a requirement of Industry 4.0, the digitalization of the models of production and processes makes it very important to ensure that the dimensions of newly manufactured parts meet the specifications of the models. Therefore, it is possible to reduce the scrap and the cost of non-conformities, ensuring the stability of the production at the same time. To ensure the quality of manufactured parts, it becomes necessary to carry out traceable measurements at scales lower than one millimeter. Providing adequate traceability to the SI unit of length (the meter) to 2D and 3D measurements at this scale is a problem that does not have a unique solution in industrial environments. Researchers in the field of dimensional metrology all around the world are working on this issue. A solution for industrial environments, even if it is not complete, will enable working with some traceability. At this point, we believe that the study of the surfaces could provide us with a first approximation to a solution. Among the different options proposed in the literature, the areal topography methods may be the most relevant because they could be compared to those measurements performed using Coordinate Measuring Machines (CMM’s). These measuring methods give (x, y, z) coordinates for each point, expressing it in two different ways, either expressing the z coordinate as a function of x, denoting it as z(x), for each Y-axis coordinate, or as a function of the x and y coordinates, denoting it as z (x, y). Between others, optical measuring instruments, mainly microscopes, are extensively used to carry out measurements at scales lower than one millimeter because it is a non-destructive measuring method. In this paper, the authors propose a calibration procedure for the scales of optical measuring instruments, particularizing for a confocal microscope, using material standards easy to find and calibrate in metrology and quality laboratories in industrial environments. Confocal microscopes are measuring instruments capable of filtering the out-of-focus reflected light so that when it reaches the detector, it is possible to take pictures of the part of the surface that is focused. Varying and taking pictures at different Z levels of the focus, a specialized software interpolates between the different planes, and it could reconstruct the surface geometry into a 3D model. As it is easy to deduce, it is necessary to give traceability to each axis. As a complementary result, the roughness Ra parameter will be traced to the reference. Although the solution is designed for a confocal microscope, it may be used for the calibration of other optical measuring instruments by applying minor changes.

Keywords: industrial environment, confocal microscope, optical measuring instrument, traceability

Procedia PDF Downloads 135
4175 Regeneration Nature of Rumex Species Root Fragment as Affected by Desiccation

Authors: Khalid Alshallash

Abstract:

Small fragments of the roots of some Rumex species including R. obtusifolius and R. crispus have been found to regenerate readily, contributing to the severity of infestations by these very common, widespread and difficult to control perennial weeds of agricultural crops and grasslands. Their root fragments are usually created during routine agricultural practices. We found that fresh root fragments of both species containing 65-70 % of moisture, progressively lose their moisture content when desiccated under controlled growth room conditions matching summer weather of southeast England, with the greatest reduction occurring in the first 48 hours. Probability of shoot emergence and the time taken for emergence in glasshouse conditions were also reduced significantly by desiccation, with R. obtusifolius least affected up to 48-hour. However, the effects converged after 120 hours. In contrast, R. obtusifolius was significantly slower to emerge after up to 48 hours desiccation, again effects converging after longer periods, R. crispus entirely failed to emerge at 120 hours. The dry weight of emerged shoots was not significantly different between the species, until desiccated for 96 hours when R. obtusifolius was significantly reduced. At 120 hours, R. obtusifolius did not emerge. In outdoor trials, desiccation for 24 or 48 hours had less effect on emergence when planted at the soil surface or up to 10 cm of depth, compared to deeper plantings. In both species, emergence was significantly lower when desiccated fragments were planted at 15 or 20 cm. Time taken for emergence was not significantly different between the species until planted at 15 or 20 cm when R. obtusifolius was slower than R. crispus and reduced further by increasing desiccation. Similar variation in effects of increasing soil depth interacting with increasing desiccation was found in reductions in dry weight, the number of tillers and leaf area, with R obtusifolius generally but not exclusively better able to withstand more extreme trial conditions. Our findings suggest that infestations of these highly troublesome weeds may be partly controlled by appropriate agricultural practices, notably exposing cut fragments to drying environmental conditions followed by deep burial.

Keywords: regeneration, root fragment, rumex crispus, rumex obtusifolius

Procedia PDF Downloads 84
4174 Perceived Restorativeness Scale– 6: A Short Version of the Perceived Restorativeness Scale for Mixed (or Mobile) Devices

Authors: Sara Gallo, Margherita Pasini, Margherita Brondino, Daniela Raccanello, Roberto Burro, Elisa Menardo

Abstract:

Most of the studies on the ability of environments to recover people’s cognitive resources have been conducted in laboratory using simulated environments (e.g., photographs, videos, or virtual reality), based on the implicit assumption that exposure to simulated environments has the same effects of exposure to real environments. However, the technical characteristics of simulated environments, such as the dynamic or static characteristics of the stimulus, critically affect their perception. Measuring perceived restorativeness in situ rather than in laboratory could increase the validity of the obtained measurements. Personal mobile devices could be useful because they allow accessing immediately online surveys when people are directly exposed to an environment. At the same time, it becomes important to develop short and reliable measuring instruments that allow a quick assessment of the restorative qualities of the environments. One of the frequently used self-report measures to assess perceived restorativeness is the “Perceived Restorativeness Scale” (PRS) based on Attention Restoration Theory. A lot of different versions have been proposed and used according to different research purposes and needs, without studying their validity. This longitudinal study reported some preliminary validation analyses on a short version of original scale, the PRS-6, developed to be quick and mobile-friendly. It is composed of 6 items assessing fascination and being-away. 102 Italian university students participated to the study, 84% female with age ranging from 18 to 47 (M = 20.7; SD = 2.9). Data were obtained through a survey online that asked them to report their perceived restorativeness of the environment they were in (and the kind of environment) and their positive emotion (Positive and Negative Affective Schedule, PANAS) once a day for seven days. Cronbach alpha and item-total correlations were used to assess reliability and internal consistency. Confirmatory Factor Analyses (CFA) models were run to study the factorial structure (construct validity). Correlation analyses between PRS and PANAS scores were used to check discriminant validity. In the end, multigroup CFA models were used to study measurement invariance (configural, metric, scalar, strict) between different mobile devices and between day of assessment. On the whole, the PRS-6 showed good psychometric proprieties, similar to those of the original scale, and invariance across devices and days. These results suggested that the PRS-6 could be a valid alternative to assess perceived restorativeness when researchers need a brief and immediate evaluation of the recovery quality of an environment.

Keywords: restorativeness, validation, short scale development, psychometrics proprieties

Procedia PDF Downloads 234
4173 A Hebbian Neural Network Model of the Stroop Effect

Authors: Vadim Kulikov

Abstract:

The classical Stroop effect is the phenomenon that it takes more time to name the ink color of a printed word if the word denotes a conflicting color than if it denotes the same color. Over the last 80 years, there have been many variations of the experiment revealing various mechanisms behind semantic, attentional, behavioral and perceptual processing. The Stroop task is known to exhibit asymmetry. Reading the words out loud is hardly dependent on the ink color, but naming the ink color is significantly influenced by the incongruent words. This asymmetry is reversed, if instead of naming the color, one has to point at a corresponding color patch. Another debated aspects are the notions of automaticity and how much of the effect is due to semantic and how much due to response stage interference. Is automaticity a continuous or an all-or-none phenomenon? There are many models and theories in the literature tackling these questions which will be discussed in the presentation. None of them, however, seems to capture all the findings at once. A computational model is proposed which is based on the philosophical idea developed by the author that the mind operates as a collection of different information processing modalities such as different sensory and descriptive modalities, which produce emergent phenomena through mutual interaction and coherence. This is the framework theory where ‘framework’ attempts to generalize the concepts of modality, perspective and ‘point of view’. The architecture of this computational model consists of blocks of neurons, each block corresponding to one framework. In the simplest case there are four: visual color processing, text reading, speech production and attention selection modalities. In experiments where button pressing or pointing is required, a corresponding block is added. In the beginning, the weights of the neural connections are mostly set to zero. The network is trained using Hebbian learning to establish connections (corresponding to ‘coherence’ in framework theory) between these different modalities. The amount of data fed into the network is supposed to mimic the amount of practice a human encounters, in particular it is assumed that converting written text into spoken words is a more practiced skill than converting visually perceived colors to spoken color-names. After the training, the network performs the Stroop task. The RT’s are measured in a canonical way, as these are continuous time recurrent neural networks (CTRNN). The above-described aspects of the Stroop phenomenon along with many others are replicated. The model is similar to some existing connectionist models but as will be discussed in the presentation, has many advantages: it predicts more data, the architecture is simpler and biologically more plausible.

Keywords: connectionism, Hebbian learning, artificial neural networks, philosophy of mind, Stroop

Procedia PDF Downloads 253
4172 Tracing Sources of Sediment in an Arid River, Southern Iran

Authors: Hesam Gholami

Abstract:

Elevated suspended sediment loads in riverine systems resulting from accelerated erosion due to human activities are a serious threat to the sustainable management of watersheds and ecosystem services therein worldwide. Therefore, mitigation of deleterious sediment effects as a distributed or non-point pollution source in the catchments requires reliable provenance information. Sediment tracing or sediment fingerprinting, as a combined process consisting of sampling, laboratory measurements, different statistical tests, and the application of mixing or unmixing models, is a useful technique for discriminating the sources of sediments. From 1996 to the present, different aspects of this technique, such as grouping the sources (spatial and individual sources), discriminating the potential sources by different statistical techniques, and modification of mixing and unmixing models, have been introduced and modified by many researchers worldwide, and have been applied to identify the provenance of fine materials in agricultural, rural, mountainous, and coastal catchments, and in large catchments with numerous lakes and reservoirs. In the last two decades, efforts exploring the uncertainties associated with sediment fingerprinting results have attracted increasing attention. The frameworks used to quantify the uncertainty associated with fingerprinting estimates can be divided into three groups comprising Monte Carlo simulation, Bayesian approaches and generalized likelihood uncertainty estimation (GLUE). Given the above background, the primary goal of this study was to apply geochemical fingerprinting within the GLUE framework in the estimation of sub-basin spatial sediment source contributions in the arid Mehran River catchment in southern Iran, which drains into the Persian Gulf. The accuracy of GLUE predictions generated using four different sets of statistical tests for discriminating three sub-basin spatial sources was evaluated using 10 virtual sediments (VS) samples with known source contributions using the root mean square error (RMSE) and mean absolute error (MAE). Based on the results, the contributions modeled by GLUE for the western, central and eastern sub-basins are 1-42% (overall mean 20%), 0.5-30% (overall mean 12%) and 55-84% (overall mean 68%), respectively. According to the mean absolute fit (MAF; ≥ 95% for all target sediment samples) and goodness-of-fit (GOF; ≥ 99% for all samples), our suggested modeling approach is an accurate technique to quantify the source of sediments in the catchments. Overall, the estimated source proportions can help watershed engineers plan the targeting of conservation programs for soil and water resources.

Keywords: sediment source tracing, generalized likelihood uncertainty estimation, virtual sediment mixtures, Iran

Procedia PDF Downloads 64
4171 Frequency Response of Complex Systems with Localized Nonlinearities

Authors: E. Menga, S. Hernandez

Abstract:

Finite Element Models (FEMs) are widely used in order to study and predict the dynamic properties of structures and usually, the prediction can be obtained with much more accuracy in the case of a single component than in the case of assemblies. Especially for structural dynamics studies, in the low and middle frequency range, most complex FEMs can be seen as assemblies made by linear components joined together at interfaces. From a modelling and computational point of view, these types of joints can be seen as localized sources of stiffness and damping and can be modelled as lumped spring/damper elements, most of time, characterized by nonlinear constitutive laws. On the other side, most of FE programs are able to run nonlinear analysis in time-domain. They treat the whole structure as nonlinear, even if there is one nonlinear degree of freedom (DOF) out of thousands of linear ones, making the analysis unnecessarily expensive from a computational point of view. In this work, a methodology in order to obtain the nonlinear frequency response of structures, whose nonlinearities can be considered as localized sources, is presented. The work extends the well-known Structural Dynamic Modification Method (SDMM) to a nonlinear set of modifications, and allows getting the Nonlinear Frequency Response Functions (NLFRFs), through an ‘updating’ process of the Linear Frequency Response Functions (LFRFs). A brief summary of the analytical concepts is given, starting from the linear formulation and understanding what the implications of the nonlinear one, are. The response of the system is formulated in both: time and frequency domain. First the Modal Database is extracted and the linear response is calculated. Secondly the nonlinear response is obtained thru the NL SDMM, by updating the underlying linear behavior of the system. The methodology, implemented in MATLAB, has been successfully applied to estimate the nonlinear frequency response of two systems. The first one is a two DOFs spring-mass-damper system, and the second example takes into account a full aircraft FE Model. In spite of the different levels of complexity, both examples show the reliability and effectiveness of the method. The results highlight a feasible and robust procedure, which allows a quick estimation of the effect of localized nonlinearities on the dynamic behavior. The method is particularly powerful when most of the FE Model can be considered as acting linearly and the nonlinear behavior is restricted to few degrees of freedom. The procedure is very attractive from a computational point of view because the FEM needs to be run just once, which allows faster nonlinear sensitivity analysis and easier implementation of optimization procedures for the calibration of nonlinear models.

Keywords: frequency response, nonlinear dynamics, structural dynamic modification, softening effect, rubber

Procedia PDF Downloads 259
4170 Analysing Competitive Advantage of IoT and Data Analytics in Smart City Context

Authors: Petra Hofmann, Dana Koniel, Jussi Luukkanen, Walter Nieminen, Lea Hannola, Ilkka Donoghue

Abstract:

The Covid-19 pandemic forced people to isolate and become physically less connected. The pandemic has not only reshaped people’s behaviours and needs but also accelerated digital transformation (DT). DT of cities has become an imperative with the outlook of converting them into smart cities in the future. Embedding digital infrastructure and smart city initiatives as part of normal design, construction, and operation of cities provides a unique opportunity to improve the connection between people. The Internet of Things (IoT) is an emerging technology and one of the drivers in DT. It has disrupted many industries by introducing different services and business models, and IoT solutions are being applied in multiple fields, including smart cities. As IoT and data are fundamentally linked together, IoT solutions can only create value if the data generated by the IoT devices is analysed properly. Extracting relevant conclusions and actionable insights by using established techniques, data analytics contributes significantly to the growth and success of IoT applications and investments. Companies must grasp DT and be prepared to redesign their offerings and business models to remain competitive in today’s marketplace. As there are many IoT solutions available today, the amount of data is tremendous. The challenge for companies is to understand what solutions to focus on and how to prioritise and which data to differentiate from the competition. This paper explains how IoT and data analytics can impact competitive advantage and how companies should approach IoT and data analytics to translate them into concrete offerings and solutions in the smart city context. The study was carried out as a qualitative, literature-based research. A case study is provided to validate the preservation of company’s competitive advantage through smart city solutions. The results of the research contribution provide insights into the different factors and considerations related to creating competitive advantage through IoT and data analytics deployment in the smart city context. Furthermore, this paper proposes a framework that merges the factors and considerations with examples of offerings and solutions in smart cities. The data collected through IoT devices, and the intelligent use of it, can create competitive advantage to companies operating in smart city business. Companies should take into consideration the five forces of competition that shape industries and pay attention to the technological, organisational, and external contexts which define factors for consideration of competitive advantages in the field of IoT and data analytics. Companies that can utilise these key assets in their businesses will most likely conquer the markets and have a strong foothold in the smart city business.

Keywords: data analytics, smart cities, competitive advantage, internet of things

Procedia PDF Downloads 118
4169 An E-Retailing System Architecture Based on Cloud Computing

Authors: Chanchai Supaartagorn

Abstract:

E-retailing is the sale of goods online that takes place over the Internet. The Internet has shrunk the entire World. The world e-retailing is growing at an exponential rate in the Americas, Europe, and Asia. However, e-retailing costs require expensive investment, such as hardware, software, and security systems. Cloud computing technology is internet-based computing for the management and delivery of applications and services. Cloud-based e-retailing application models allow enterprises to lower their costs with their effective implementation of e-retailing activities. In this paper, we describe the concept of cloud computing and present the architecture of cloud computing, combining the features of e-retailing. In addition, we propose a strategy for implementing cloud computing with e-retailing. Finally, we explain the benefits from the architecture.

Keywords: architecture, cloud computing, e-retailing, internet-based

Procedia PDF Downloads 383
4168 Nurse-Patient Assignment: Case of Pediatrics Department

Authors: Jihene Jlassi, Ahmed Frikha, Wazna Kortli

Abstract:

The objectives of Nurse-Patient Assignment are the minimization of the overall hospital cost and the maximization of nurses ‘preferences. This paper aims to assess nurses' satisfaction related to the implementation of patient acuity tool-based assignments. So, we used an integer linear program that assigns patients to nurses while balancing nurse workloads. Then, the proposed model is applied to the Paediatrics Department at Kasserine Hospital Tunisia. Where patients need special acuities and high-level nursing skills and care. Hence, numerical results suggested that proposed nurse-patient assignment models can achieve a balanced assignment

Keywords: nurse-patient assignment, mathematical model, logistics, pediatrics department, balanced assignment

Procedia PDF Downloads 137
4167 Real-time Rate and Rhythms Feedback Control System in Patients with Atrial Fibrillation

Authors: Mohammad A. Obeidat, Ayman M. Mansour

Abstract:

Capturing the dynamic behavior of the heart to improve control performance, enhance robustness, and support diagnosis is very important in establishing real time models for the heart. Control Techniques and strategies have been utilized to improve system costs, reliability, and estimation accuracy for different types of systems such as biomedical, industrial, and other systems that required tuning input/output relation and/or monitoring. Simulations are performed to illustrate potential applications of the technology. In this research, a new control technology scheme is used to enhance the performance of the Af system and meet the design specifications.

Keywords: atrial fibrillation, dynamic behavior, closed loop, signal, filter

Procedia PDF Downloads 401
4166 Discourse Analysis: Where Cognition Meets Communication

Authors: Iryna Biskub

Abstract:

The interdisciplinary approach to modern linguistic studies is exemplified by the merge of various research methods, which sometimes causes complications related to the verification of the research results. This methodological confusion can be resolved by means of creating new techniques of linguistic analysis combining several scientific paradigms. Modern linguistics has developed really productive and efficient methods for the investigation of cognitive and communicative phenomena of which language is the central issue. In the field of discourse studies, one of the best examples of research methods is the method of Critical Discourse Analysis (CDA). CDA can be viewed both as a method of investigation, as well as a critical multidisciplinary perspective. In CDA the position of the scholar is crucial from the point of view exemplifying his or her social and political convictions. The generally accepted approach to obtaining scientifically reliable results is to use a special well-defined scientific method for researching special types of language phenomena: cognitive methods applied to the exploration of cognitive aspects of language, whereas communicative methods are thought to be relevant only for the investigation of communicative nature of language. In the recent decades discourse as a sociocultural phenomenon has been the focus of careful linguistic research. The very concept of discourse represents an integral unity of cognitive and communicative aspects of human verbal activity. Since a human being is never able to discriminate between cognitive and communicative planes of discourse communication, it doesn’t make much sense to apply cognitive and communicative methods of research taken in isolation. It is possible to modify the classical CDA procedure by means of mapping human cognitive procedures onto the strategic communicative planning of discourse communication. The analysis of the electronic petition 'Block Donald J Trump from UK entry. The signatories believe Donald J Trump should be banned from UK entry' (584, 459 signatures) and the parliamentary debates on it has demonstrated the ability to map cognitive and communicative levels in the following way: the strategy of discourse modeling (communicative level) overlaps with the extraction of semantic macrostructures (cognitive level); the strategy of discourse management overlaps with the analysis of local meanings in discourse communication; the strategy of cognitive monitoring of the discourse overlaps with the formation of attitudes and ideologies at the cognitive level. Thus, the experimental data have shown that it is possible to develop a new complex methodology of discourse analysis, where cognition would meet communication, both metaphorically and literally. The same approach may appear to be productive for the creation of computational models of human-computer interaction, where the automatic generation of a particular type of a discourse could be based on the rules of strategic planning involving cognitive models of CDA.

Keywords: cognition, communication, discourse, strategy

Procedia PDF Downloads 242