Search results for: hardy cross networks accuracy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9694

Search results for: hardy cross networks accuracy

6244 Biimodal Biometrics System Using Fusion of Iris and Fingerprint

Authors: Attallah Bilal, Hendel Fatiha

Abstract:

This paper proposes the bimodal biometrics system for identity verification iris and fingerprint, at matching score level architecture using weighted sum of score technique. The features are extracted from the pre processed images of iris and fingerprint. These features of a query image are compared with those of a database image to obtain matching scores. The individual scores generated after matching are passed to the fusion module. This module consists of three major steps i.e., normalization, generation of similarity score and fusion of weighted scores. The final score is then used to declare the person as genuine or an impostor. The system is tested on CASIA database and gives an overall accuracy of 91.04% with FAR of 2.58% and FRR of 8.34%.

Keywords: iris, fingerprint, sum rule, fusion

Procedia PDF Downloads 363
6243 Model Predictive Control of Three Phase Inverter for PV Systems

Authors: Irtaza M. Syed, Kaamran Raahemifar

Abstract:

This paper presents a model predictive control (MPC) of a utility interactive three phase inverter (TPI) for a photovoltaic (PV) system at commercial level. The proposed model uses phase locked loop (PLL) to synchronize TPI with the power electric grid (PEG) and performs MPC control in a dq reference frame. TPI model consists of boost converter (BC), maximum power point tracking (MPPT) control, and a three leg voltage source inverter (VSI). Operational model of VSI is used to synthesize sinusoidal current and track the reference. Model is validated using a 35.7 kW PV system in Matlab/Simulink. Implementation and results show simplicity and accuracy, as well as reliability of the model.

Keywords: model predictive control, three phase voltage source inverter, PV system, Matlab/simulink

Procedia PDF Downloads 586
6242 Compensatory Neuro-Fuzzy Inference (CNFI) Controller for Bilateral Teleoperation

Authors: R. Mellah, R. Toumi

Abstract:

This paper presents a new adaptive neuro-fuzzy controller equipped with compensatory fuzzy control (CNFI) in order to not only adjusts membership functions but also to optimize the adaptive reasoning by using a compensatory learning algorithm. The proposed control structure includes both CNFI controllers for which one is used to control in force the master robot and the second one for controlling in position the slave robot. The experimental results obtained, show a fairly high accuracy in terms of position and force tracking under free space motion and hard contact motion, what highlights the effectiveness of the proposed controllers.

Keywords: compensatory fuzzy, neuro-fuzzy, control adaptive, teleoperation

Procedia PDF Downloads 320
6241 Real-time Rate and Rhythms Feedback Control System in Patients with Atrial Fibrillation

Authors: Mohammad A. Obeidat, Ayman M. Mansour

Abstract:

Capturing the dynamic behavior of the heart to improve control performance, enhance robustness, and support diagnosis is very important in establishing real time models for the heart. Control Techniques and strategies have been utilized to improve system costs, reliability, and estimation accuracy for different types of systems such as biomedical, industrial, and other systems that required tuning input/output relation and/or monitoring. Simulations are performed to illustrate potential applications of the technology. In this research, a new control technology scheme is used to enhance the performance of the Af system and meet the design specifications.

Keywords: atrial fibrillation, dynamic behavior, closed loop, signal, filter

Procedia PDF Downloads 409
6240 Numerical Analysis of Liquid Metal Magnetohydrodynamic Flows in a Manifold with Three Sub-Channels

Authors: Meimei Wen, Chang Nyung Kim

Abstract:

In the current study, three-dimensional liquid metal (LM) magneto-hydrodynamic (MHD) flows in a manifold with three sub-channels under a uniform magnetic field are numerically investigated. In the manifold, the electrical current can cross channel walls, thus having influence on the flow distribution in each sub-channel. A case with various arrangements of electric conductivity for different parts of channel walls is considered, yielding different current distributions as well as flow distributions in each sub-channel. Here, the imbalance of mass flow rates in the three sub-channels is addressed. Meanwhile, predicted are detailed behaviors of the flow velocity, pressure, current and electric potential of LM MHD flows with three sub-channels. Commercial software CFX is used for the numerical simulation of LM MHD flows.

Keywords: CFX, liquid metal, manifold, MHD flow

Procedia PDF Downloads 342
6239 Unveiling Comorbidities in Irritable Bowel Syndrome: A UK BioBank Study utilizing Supervised Machine Learning

Authors: Uswah Ahmad Khan, Muhammad Moazam Fraz, Humayoon Shafique Satti, Qasim Aziz

Abstract:

Approximately 10-14% of the global population experiences a functional disorder known as irritable bowel syndrome (IBS). The disorder is defined by persistent abdominal pain and an irregular bowel pattern. IBS significantly impairs work productivity and disrupts patients' daily lives and activities. Although IBS is widespread, there is still an incomplete understanding of its underlying pathophysiology. This study aims to help characterize the phenotype of IBS patients by differentiating the comorbidities found in IBS patients from those in non-IBS patients using machine learning algorithms. In this study, we extracted samples coding for IBS from the UK BioBank cohort and randomly selected patients without a code for IBS to create a total sample size of 18,000. We selected the codes for comorbidities of these cases from 2 years before and after their IBS diagnosis and compared them to the comorbidities in the non-IBS cohort. Machine learning models, including Decision Trees, Gradient Boosting, Support Vector Machine (SVM), AdaBoost, Logistic Regression, and XGBoost, were employed to assess their accuracy in predicting IBS. The most accurate model was then chosen to identify the features associated with IBS. In our case, we used XGBoost feature importance as a feature selection method. We applied different models to the top 10% of features, which numbered 50. Gradient Boosting, Logistic Regression and XGBoost algorithms yielded a diagnosis of IBS with an optimal accuracy of 71.08%, 71.427%, and 71.53%, respectively. Among the comorbidities most closely associated with IBS included gut diseases (Haemorrhoids, diverticular diseases), atopic conditions(asthma), and psychiatric comorbidities (depressive episodes or disorder, anxiety). This finding emphasizes the need for a comprehensive approach when evaluating the phenotype of IBS, suggesting the possibility of identifying new subsets of IBS rather than relying solely on the conventional classification based on stool type. Additionally, our study demonstrates the potential of machine learning algorithms in predicting the development of IBS based on comorbidities, which may enhance diagnosis and facilitate better management of modifiable risk factors for IBS. Further research is necessary to confirm our findings and establish cause and effect. Alternative feature selection methods and even larger and more diverse datasets may lead to more accurate classification models. Despite these limitations, our findings highlight the effectiveness of Logistic Regression and XGBoost in predicting IBS diagnosis.

Keywords: comorbidities, disease association, irritable bowel syndrome (IBS), predictive analytics

Procedia PDF Downloads 112
6238 Computational and Experimental Determination of Acoustic Impedance of Internal Combustion Engine Exhaust

Authors: A. O. Glazkov, A. S. Krylova, G. G. Nadareishvili, A. S. Terenchenko, S. I. Yudin

Abstract:

The topic of the presented materials concerns the design of the exhaust system for a certain internal combustion engine. The exhaust system can be divided into two parts. The first is the engine exhaust manifold, turbocharger, and catalytic converters, which are called “hot part.” The second part is the gas exhaust system, which contains elements exclusively for reducing exhaust noise (mufflers, resonators), the accepted designation of which is the "cold part." The design of the exhaust system from the point of view of acoustics, that is, reducing the exhaust noise to a predetermined level, consists of working on the second part. Modern computer technology and software make it possible to design "cold part" with high accuracy in a given frequency range but with the condition of accurately specifying the input parameters, namely, the amplitude spectrum of the input noise and the acoustic impedance of the noise source in the form of an engine with a "hot part". Getting this data is a difficult problem: high temperatures, high exhaust gas velocities (turbulent flows), and high sound pressure levels (non-linearity mode) do not allow the calculated results to be applied with sufficient accuracy. The aim of this work is to obtain the most reliable acoustic output parameters of an engine with a "hot part" based on a complex of computational and experimental studies. The presented methodology includes several parts. The first part is a finite element simulation of the "cold part" of the exhaust system (taking into account the acoustic impedance of radiation of outlet pipe into open space) with the result in the form of the input impedance of "cold part". The second part is a finite element simulation of the "hot part" of the exhaust system (taking into account acoustic characteristics of catalytic units and geometry of turbocharger) with the result in the form of the input impedance of the "hot part". The next third part of the technique consists of the mathematical processing of the results according to the proposed formula for the convergence of the mathematical series of summation of multiple reflections of the acoustic signal "cold part" - "hot part". This is followed by conducting a set of tests on an engine stand with two high-temperature pressure sensors measuring pulsations in the nozzle between "hot part" and "cold part" of the exhaust system and subsequent processing of test results according to a well-known technique in order to separate the "incident" and "reflected" waves. The final stage consists of the mathematical processing of all calculated and experimental data to obtain a result in the form of a spectrum of the amplitude of the engine noise and its acoustic impedance.

Keywords: acoustic impedance, engine exhaust system, FEM model, test stand

Procedia PDF Downloads 54
6237 Size Reduction of Images Using Constraint Optimization Approach for Machine Communications

Authors: Chee Sun Won

Abstract:

This paper presents the size reduction of images for machine-to-machine communications. Here, the salient image regions to be preserved include the image patches of the key-points such as corners and blobs. Based on a saliency image map from the key-points and their image patches, an axis-aligned grid-size optimization is proposed for the reduction of image size. To increase the size-reduction efficiency the aspect ratio constraint is relaxed in the constraint optimization framework. The proposed method yields higher matching accuracy after the size reduction than the conventional content-aware image size-reduction methods.

Keywords: image compression, image matching, key-point detection and description, machine-to-machine communication

Procedia PDF Downloads 414
6236 Comparative Study of Ad Hoc Routing Protocols in Vehicular Ad-Hoc Networks for Smart City

Authors: Khadija Raissi, Bechir Ben Gouissem

Abstract:

In this paper, we perform the investigation of some routing protocols in Vehicular Ad-Hoc Network (VANET) context. Indeed, we study the efficiency of protocols like Dynamic Source Routing (DSR), Ad hoc On-demand Distance Vector Routing (AODV), Destination Sequenced Distance Vector (DSDV), Optimized Link State Routing convention (OLSR) and Vehicular Multi-hop algorithm for Stable Clustering (VMASC) in terms of packet delivery ratio (PDR) and throughput. The performance evaluation and comparison between the studied protocols shows that the VMASC is the best protocols regarding fast data transmission and link stability in VANETs. The validation of all results is done by the NS3 simulator.

Keywords: VANET, smart city, AODV, OLSR, DSR, OLSR, VMASC, routing protocols, NS3

Procedia PDF Downloads 288
6235 Theorizing Women’s Political Leadership: Cross-National Comparison

Authors: Minjeoung Kim

Abstract:

Since women obtained the right to vote in 1893 for the first time in New Zealand, they have tried to participate actively into politics but still the world has a few women in political leadership. The article asks which factors might influence the appearance of women leadership in politics. The article investigates two factors such as political context, personal factors. Countries where economic development is stable and political democracy is consolidated have a tendency of appearance of women political leadership but in less developed and politically unstable countries, women politicians can be in power with their own reasons. For the personal factor, their feminist propensity is studied but there is no relationship between the appearance of women leaders and their feminist propensity.

Keywords: women political leadership, political context, slow track, transitory countries, feminist propensity

Procedia PDF Downloads 403
6234 Rapid Soil Classification Using Computer Vision with Electrical Resistivity and Soil Strength

Authors: Eugene Y. J. Aw, J. W. Koh, S. H. Chew, K. E. Chua, P. L. Goh, Grace H. B. Foo, M. L. Leong

Abstract:

This paper presents the evaluation of various soil testing methods such as the four-probe soil electrical resistivity method and cone penetration test (CPT) that can complement a newly developed novel rapid soil classification scheme using computer vision, to improve the accuracy and productivity of on-site classification of excavated soil. In Singapore, excavated soils from the local construction industry are transported to Staging Grounds (SGs) to be reused as fill material for land reclamation. Excavated soils are mainly categorized into two groups (“Good Earth” and “Soft Clay”) based on particle size distribution (PSD) and water content (w) from soil investigation reports and on-site visual survey, such that proper treatment and usage can be exercised. However, this process is time-consuming and labor-intensive. Thus, a rapid classification method is needed at the SGs. Four-probe soil electrical resistivity and CPT were evaluated for their feasibility as suitable additions to the computer vision system to further develop this innovative non-destructive and instantaneous classification method. The computer vision technique comprises soil image acquisition using an industrial-grade camera; image processing and analysis via calculation of Grey Level Co-occurrence Matrix (GLCM) textural parameters; and decision-making using an Artificial Neural Network (ANN). It was found from the previous study that the ANN model coupled with ρ can classify soils into “Good Earth” and “Soft Clay” in less than a minute, with an accuracy of 85% based on selected representative soil images. To further improve the technique, the following three items were targeted to be added onto the computer vision scheme: the apparent electrical resistivity of soil (ρ) measured using a set of four probes arranged in Wenner’s array, the soil strength measured using a modified mini cone penetrometer, and w measured using a set of time-domain reflectometry (TDR) probes. Laboratory proof-of-concept was conducted through a series of seven tests with three types of soils – “Good Earth”, “Soft Clay,” and a mix of the two. Validation was performed against the PSD and w of each soil type obtained from conventional laboratory tests. The results show that ρ, w and CPT measurements can be collectively analyzed to classify soils into “Good Earth” or “Soft Clay” and are feasible as complementing methods to the computer vision system.

Keywords: computer vision technique, cone penetration test, electrical resistivity, rapid and non-destructive, soil classification

Procedia PDF Downloads 230
6233 [Keynote Talk]: The Intoxicated Eyewitness: Effect of Alcohol Consumption on Identification Accuracy in Lineup

Authors: Vikas S. Minchekar

Abstract:

The eyewitness is a crucial source of evidence in the criminal judicial system. However, rely on the reminiscence of an eyewitness especially intoxicated eyewitness is not always judicious. It might lead to some serious consequences. Day by day, alcohol-related crimes or the criminal incidences in bars, nightclubs, and restaurants are increasing rapidly. Tackling such cases is very complicated to any investigation officers. The people in that incidents are violated due to the alcohol consumption hence, their ability to identify the suspects or recall these phenomena is affected. The studies on the effects of alcohol consumption on motor activities such as driving and surgeries have received much attention. However, the effect of alcohol intoxication on memory has received little attention from the psychology, law, forensic and criminology scholars across the world. In the Indian context, the published articles on this issue are equal to none up to present day. This field experiment investigation aimed at to finding out the effect of alcohol consumption on identification accuracy in lineups. Forty adult, social drinkers, and twenty sober adults were randomly recruited for the study. The sober adults were assigned into 'placebo' beverage group while social drinkers were divided into two group e. g. 'low dose' of alcohol (0.2 g/kg) and 'high dose' of alcohol (0.8 g/kg). The social drinkers were divided in such a way that their level of blood-alcohol concentration (BAC) will become different. After administering the beverages for the placebo group and liquor to the social drinkers for 40 to 50 minutes of the period, the five-minute video clip of mock crime is shown to all in a group of four to five members. After the exposure of video, clip subjects were given 10 portraits and asked them to recognize whether they are involved in mock crime or not. Moreover, they were also asked to describe the incident. The subjects were given two opportunities to recognize the portraits and to describe the events; the first opportunity is given immediately after the video clip and the second was 24 hours later. The obtained data were analyzed by one-way ANOVA and Scheffe’s posthoc multiple comparison tests. The results indicated that the 'high dose' group is remarkably different from the 'placebo' and 'low dose' groups. But, the 'placebo' and 'low dose' groups are equally performed. The subjects in a 'high dose' group recognized only 20% faces correctly while the subjects in a 'placebo' and 'low dose' groups are recognized 90 %. This study implied that the intoxicated witnesses are less accurate to recognize the suspects and also less capable of describing the incidents where crime has taken place. Moreover, this study does not assert that intoxicated eyewitness is generally less trustworthy than their sober counterparts.

Keywords: intoxicated eyewitness, memory, social drinkers, lineups

Procedia PDF Downloads 265
6232 A Study on Numerical Modelling of Rigid Pavement: Temperature and Thickness Effect

Authors: Amin Chegenizadeh, Mahdi Keramatikerman, Hamid Nikraz

Abstract:

Pavement engineering plays a significant role to develop cost effective and efficient highway and road networks. In general, pavement regarding structure is categorized in two core group namely flexible and rigid pavements. There are various benefits in application of rigid pavement. For instance, they have a longer life and lower maintenance costs in compare with the flexible pavement. In rigid pavement designs, temperature and thickness are two effective parameters that could widely affect the total cost of the project. In this study, a numerical modeling using Kenpave-Kenslab was performed to investigate the effect of these two important parameters in the rigid pavement.   

Keywords: rigid pavement, Kenpave, Kenslab, thickness, temperature

Procedia PDF Downloads 366
6231 Extraction of Urban Building Damage Using Spectral, Height and Corner Information

Authors: X. Wang

Abstract:

Timely and accurate information on urban building damage caused by earthquake is important basis for disaster assessment and emergency relief. Very high resolution (VHR) remotely sensed imagery containing abundant fine-scale information offers a large quantity of data for detecting and assessing urban building damage in the aftermath of earthquake disasters. However, the accuracy obtained using spectral features alone is comparatively low, since building damage, intact buildings and pavements are spectrally similar. Therefore, it is of great significance to detect urban building damage effectively using multi-source data. Considering that in general height or geometric structure of buildings change dramatically in the devastated areas, a novel multi-stage urban building damage detection method, using bi-temporal spectral, height and corner information, was proposed in this study. The pre-event height information was generated using stereo VHR images acquired from two different satellites, while the post-event height information was produced from airborne LiDAR data. The corner information was extracted from pre- and post-event panchromatic images. The proposed method can be summarized as follows. To reduce the classification errors caused by spectral similarity and errors in extracting height information, ground surface, shadows, and vegetation were first extracted using the post-event VHR image and height data and were masked out. Two different types of building damage were then extracted from the remaining areas: the height difference between pre- and post-event was used for detecting building damage showing significant height change; the difference in the density of corners between pre- and post-event was used for extracting building damage showing drastic change in geometric structure. The initial building damage result was generated by combining above two building damage results. Finally, a post-processing procedure was adopted to refine the obtained initial result. The proposed method was quantitatively evaluated and compared to two existing methods in Port au Prince, Haiti, which was heavily hit by an earthquake in January 2010, using pre-event GeoEye-1 image, pre-event WorldView-2 image, post-event QuickBird image and post-event LiDAR data. The results showed that the method proposed in this study significantly outperformed the two comparative methods in terms of urban building damage extraction accuracy. The proposed method provides a fast and reliable method to detect urban building collapse, which is also applicable to relevant applications.

Keywords: building damage, corner, earthquake, height, very high resolution (VHR)

Procedia PDF Downloads 210
6230 Smart Structures for Cost Effective Cultural Heritage Preservation

Authors: Tamara Trček Pečak, Andrej Mohar, Denis Trček

Abstract:

This article investigates the latest technological means, which deploy smart structures that are based on (advanced) wireless sensors technologies and ubiquitous computing in general in order to support the above mentioned decision making. Based on two years of in-field research experiences it gives their analysis for these kinds of purposes and provides appropriate architectures and architectural solutions. Moreover, the directions for future research are stated, because these technologies are currently the most promising ones to enable cost-effective preservation of cultural heritage not only in uncontrolled places, but also in general.

Keywords: smart structures, wireless sensors, sensors networks, green computing, cultural heritage preservation, monitoring, cost effectiveness

Procedia PDF Downloads 440
6229 DTI Connectome Changes in the Acute Phase of Aneurysmal Subarachnoid Hemorrhage Improve Outcome Classification

Authors: Sarah E. Nelson, Casey Weiner, Alexander Sigmon, Jun Hua, Haris I. Sair, Jose I. Suarez, Robert D. Stevens

Abstract:

Graph-theoretical information from structural connectomes indicated significant connectivity changes and improved acute prognostication in a Random Forest (RF) model in aneurysmal subarachnoid hemorrhage (aSAH), which can lead to significant morbidity and mortality and has traditionally been fraught by poor methods to predict outcome. This study’s hypothesis was that structural connectivity changes occur in canonical brain networks of acute aSAH patients, and that these changes are associated with functional outcome at six months. In a prospective cohort of patients admitted to a single institution for management of acute aSAH, patients underwent diffusion tensor imaging (DTI) as part of a multimodal MRI scan. A weighted undirected structural connectome was created of each patient’s images using Constant Solid Angle (CSA) tractography, with 176 regions of interest (ROIs) defined by the Johns Hopkins Eve atlas. ROIs were sorted into four networks: Default Mode Network, Executive Control Network, Salience Network, and Whole Brain. The resulting nodes and edges were characterized using graph-theoretic features, including Node Strength (NS), Betweenness Centrality (BC), Network Degree (ND), and Connectedness (C). Clinical (including demographics and World Federation of Neurologic Surgeons scale) and graph features were used separately and in combination to train RF and Logistic Regression classifiers to predict two outcomes: dichotomized modified Rankin Score (mRS) at discharge and at six months after discharge (favorable outcome mRS 0-2, unfavorable outcome mRS 3-6). A total of 56 aSAH patients underwent DTI a median (IQR) of 7 (IQR=8.5) days after admission. The best performing model (RF) combining clinical and DTI graph features had a mean Area Under the Receiver Operator Characteristic Curve (AUROC) of 0.88 ± 0.00 and Area Under the Precision Recall Curve (AUPRC) of 0.95 ± 0.00 over 500 trials. The combined model performed better than the clinical model alone (AUROC 0.81 ± 0.01, AUPRC 0.91 ± 0.00). The highest-ranked graph features for prediction were NS, BC, and ND. These results indicate reorganization of the connectome early after aSAH. The performance of clinical prognostic models was increased significantly by the inclusion of DTI-derived graph connectivity metrics. This methodology could significantly improve prognostication of aSAH.

Keywords: connectomics, diffusion tensor imaging, graph theory, machine learning, subarachnoid hemorrhage

Procedia PDF Downloads 182
6228 Consolidated Predictive Model of the Natural History of Breast Cancer Considering Primary Tumor and Secondary Distant Metastases Growth

Authors: Ella Tyuryumina, Alexey Neznanov

Abstract:

This study is an attempt to obtain reliable data on the natural history of breast cancer growth. We analyze the opportunities for using classical mathematical models (exponential and logistic tumor growth models, Gompertz and von Bertalanffy tumor growth models) to try to describe growth of the primary tumor and the secondary distant metastases of human breast cancer. The research aim is to improve predicting accuracy of breast cancer progression using an original mathematical model referred to CoMPaS and corresponding software. We are interested in: 1) modelling the whole natural history of the primary tumor and the secondary distant metastases; 2) developing adequate and precise CoMPaS which reflects relations between the primary tumor and the secondary distant metastases; 3) analyzing the CoMPaS scope of application; 4) implementing the model as a software tool. The foundation of the CoMPaS is the exponential tumor growth model, which is described by determinate nonlinear and linear equations. The CoMPaS corresponds to TNM classification. It allows to calculate different growth periods of the primary tumor and the secondary distant metastases: 1) ‘non-visible period’ for the primary tumor; 2) ‘non-visible period’ for the secondary distant metastases; 3) ‘visible period’ for the secondary distant metastases. The CoMPaS is validated on clinical data of 10-years and 15-years survival depending on the tumor stage and diameter of the primary tumor. The new predictive tool: 1) is a solid foundation to develop future studies of breast cancer growth models; 2) does not require any expensive diagnostic tests; 3) is the first predictor which makes forecast using only current patient data, the others are based on the additional statistical data. The CoMPaS model and predictive software: a) fit to clinical trials data; b) detect different growth periods of the primary tumor and the secondary distant metastases; c) make forecast of the period of the secondary distant metastases appearance; d) have higher average prediction accuracy than the other tools; e) can improve forecasts on survival of breast cancer and facilitate optimization of diagnostic tests. The following are calculated by CoMPaS: the number of doublings for ‘non-visible’ and ‘visible’ growth period of the secondary distant metastases; tumor volume doubling time (days) for ‘non-visible’ and ‘visible’ growth period of the secondary distant metastases. The CoMPaS enables, for the first time, to predict ‘whole natural history’ of the primary tumor and the secondary distant metastases growth on each stage (pT1, pT2, pT3, pT4) relying only on the primary tumor sizes. Summarizing: a) CoMPaS describes correctly the primary tumor growth of IA, IIA, IIB, IIIB (T1-4N0M0) stages without metastases in lymph nodes (N0); b) facilitates the understanding of the appearance period and inception of the secondary distant metastases.

Keywords: breast cancer, exponential growth model, mathematical model, metastases in lymph nodes, primary tumor, survival

Procedia PDF Downloads 336
6227 Hybrid Model of Strategic and Contextual Leadership in Pluralistic Organizations- A Qualitative Multiple Case Study

Authors: Ergham Al Bachir

Abstract:

This study adopts strategic leadership (Upper Echelons) as the core theory and contextual leadership theory as the research lens. This research asks how the external context impacts strategic leadership effectiveness to achieve the outcomes in pluralistic organizations (PO). The study explores how the context influences the selection of CEOs, top management teams (TMT), and their leadership effectiveness. POs are characterized by the multiple objectives of their top management teams, divergent objectives, multiple strategies, and multiple governing authorities. The research question is explored by means of a qualitative multiple-case study focusing on healthcare, real estate, and financial services organizations. The data sources are semi-structured interviews, documents, and direct observations. The data analysis strategy is inductive and deploys thematic analysis and cross-case synthesis. The findings differentiate between national and international CEOs' delegation of authority and relationship with the Board of Directors. The findings identify the elements of the dynamic context that influence TMT and PO outcomes. The emergent hybrid strategic and contextual leadership framework shows how the different contextual factors influence strategic direction, PO context, selection of CEOs and TMT, and the outcomes in four pluralistic organizations. The study offers seven theoretical contributions to Upper Echelons, strategic leadership, and contextual leadership research. (1) The integration of two theories revealed how CEO’s impact on the organization is complementary to the contextual impact. (2) Conducting this study in the Middle East contributes to strategic leadership and contextual leadership research. (3) The demonstration of the significant contextual effects on the selection of CEOs. (4 and 5) Two contributions revealed new links between the context, the Board role, internal versus external CEOs, and national versus international CEOs. (6 and 7) This study offered two definitions: what accounts for CEO leadership effectiveness and organizational outcomes. Two methodological contributions were also identified: (1) Previous strategic leadership and Upper Echelons research are mainly quantitative, while this study adopts qualitative multiple-case research with face-to-face interviews. (2) The extrication of the CEO from the TMT advanced the data analysis in strategic leadership research. Four contributions are offered to practice: (1) The CEO's leadership effectiveness inside and outside the organization. (2) Rapid turnover of predecessor CEOs signifies the need for a strategic and contextual approach to CEOs' succession. (3) TMT composition and education impact on TMT-CEO and TMT-TMT interface. (4) Multilevel strategic contextual leadership development framework.

Keywords: strategic leadership, contextual leadership, upper echelons, pluralistic organizations, cross-cultural leadership

Procedia PDF Downloads 82
6226 Implicature of Jokes in Broadcast Messages

Authors: Yuli Widiana

Abstract:

The study of implicature which is one of the discussions of pragmatics is an interesting and challenging topic to discuss. Implicature is a meaning which is implied in an utterance which is not the same as its literal meaning. The rapid development of information technology results in social networks as media to broadcast messages. The broadcast messages may be in the form of jokes which contain implicature. The research applies the pragmatic equivalent method to analyze the topics of jokes based on the implicatures contained in them. Furthermore, the method is also applied to reveal the purpose of creating implicature in jokes. The findings include the kinds of implicature found in jokes which are classified into conventional implicature and conversational implicature. Then, in detailed analysis, implicature in jokes is divided into implicature related to gender, culture, and social phenomena. Furthermore, implicature in jokes may not only be used to give entertainment but also to soften criticisms or satire so that it does not sound rude and harsh.

Keywords: implicature, broadcast messages, conventional implicature, conversational implicature

Procedia PDF Downloads 353
6225 Proposal for a Web System for the Control of Fungal Diseases in Grapes in Fruits Markets

Authors: Carlos Tarmeño Noriega, Igor Aguilar Alonso

Abstract:

Fungal diseases are common in vineyards; they cause a decrease in the quality of the products that can be sold, generating distrust of the customer towards the seller when buying fruit. Currently, technology allows the classification of fruits according to their characteristics thanks to artificial intelligence. This study proposes the implementation of a control system that allows the identification of the main fungal diseases present in the Italia grape, making use of a convolutional neural network (CNN), OpenCV, and TensorFlow. The methodology used was based on a collection of 20 articles referring to the proposed research on quality control, classification, and recognition of fruits through artificial vision techniques.

Keywords: computer vision, convolutional neural networks, quality control, fruit market, OpenCV, TensorFlow

Procedia PDF Downloads 77
6224 Geology and Geochemistry of the Paleozoic Basement, Western Algeria

Authors: Hadj Mohamed Nacera, Boutaleb Abdelhak

Abstract:

The Hercynian granite in Western Algeria, has a typical high-K calc-alkaline evolution, with peraluminous trend U-Pb zircon geochronology yielded the minimum emplacement age of 297 ± 1 Ma. It shows dark microgranular enclaves, veins of pegmatite, aplite, tourmaline and quartz. The granite plutons selected for this study are formed during the late Variscian phase and intrudes the Lower Silurian metasediments which were affected by the major Hercynian folding phases. An important Quartz vein field cross-cutting metasedimentary and granitic rocks. Invisible gold occurs in a very small arsenopyrite minerals. The purpose of this study is to highlight the relationship between the gold mineralisation and the intrusion by combining petrographic and geochemic studies.

Keywords: Algeria, basement, geochemestry, granite

Procedia PDF Downloads 267
6223 Time Compression in Engineer-to-Order Industry: A Case Study of a Norwegian Shipbuilding Industry

Authors: Tarek Fatouh, Chehab Elbelehy, Alaa Abdelsalam, Eman Elakkad, Alaa Abdelshafie

Abstract:

This paper aims to explore the possibility of time compression in Engineer to Order production networks. A case study research method is used in a Norwegian shipbuilding project by implementing a value stream mapping lean tool with total cycle time as a unit of analysis. The analysis resulted in demonstrating the time deviations for the planned tasks in one of the processes in the shipbuilding project. So, authors developed a future state map by removing time wastes from value stream process.

Keywords: engineer to order, total cycle time, value stream mapping, shipbuilding

Procedia PDF Downloads 157
6222 Parallel Computing: Offloading Matrix Multiplication to GPU

Authors: Bharath R., Tharun Sai N., Bhuvan G.

Abstract:

This project focuses on developing a Parallel Computing method aimed at optimizing matrix multiplication through GPU acceleration. Addressing algorithmic challenges, GPU programming intricacies, and integration issues, the project aims to enhance efficiency and scalability. The methodology involves algorithm design, GPU programming, and optimization techniques. Future plans include advanced optimizations, extended functionality, and integration with high-level frameworks. User engagement is emphasized through user-friendly interfaces, open- source collaboration, and continuous refinement based on feedback. The project's impact extends to significantly improving matrix multiplication performance in scientific computing and machine learning applications.

Keywords: matrix multiplication, parallel processing, cuda, performance boost, neural networks

Procedia PDF Downloads 49
6221 Predicting the Success of Bank Telemarketing Using Artificial Neural Network

Authors: Mokrane Selma

Abstract:

The shift towards decision making (DM) based on artificial intelligence (AI) techniques will change the way in which consumer markets and our societies function. Through AI, predictive analytics is being used by businesses to identify these patterns and major trends with the objective to improve the DM and influence future business outcomes. This paper proposes an Artificial Neural Network (ANN) approach to predict the success of telemarketing calls for selling bank long-term deposits. To validate the proposed model, we uses the bank marketing data of 41188 phone calls. The ANN attains 98.93% of accuracy which outperforms other conventional classifiers and confirms that it is credible and valuable approach for telemarketing campaign managers.

Keywords: bank telemarketing, prediction, decision making, artificial intelligence, artificial neural network

Procedia PDF Downloads 147
6220 Evaluated Nuclear Data Based Photon Induced Nuclear Reaction Model of GEANT4

Authors: Jae Won Shin

Abstract:

We develop an evaluated nuclear data based photonuclear reaction model of GEANT4 for a more accurate simulation of photon-induced neutron production. The evaluated photonuclear data libraries from the ENDF/B-VII.1 are taken as input. Incident photon energies up to 140 MeV which is the threshold energy for the pion production are considered. For checking the validity of the use of the data-based model, we calculate the photoneutron production cross-sections and yields and compared them with experimental data. The results obtained from the developed model are found to be in good agreement with the experimental data for (γ,xn) reactions.

Keywords: ENDF/B-VII.1, GEANT4, photoneutron, photonuclear reaction

Procedia PDF Downloads 271
6219 Towards an Effective Approach for Modelling near Surface Air Temperature Combining Weather and Satellite Data

Authors: Nicola Colaninno, Eugenio Morello

Abstract:

The urban environment affects local-to-global climate and, in turn, suffers global warming phenomena, with worrying impacts on human well-being, health, social and economic activities. Physic-morphological features of the built-up space affect urban air temperature, locally, causing the urban environment to be warmer compared to surrounding rural. This occurrence, typically known as the Urban Heat Island (UHI), is normally assessed by means of air temperature from fixed weather stations and/or traverse observations or based on remotely sensed Land Surface Temperatures (LST). The information provided by ground weather stations is key for assessing local air temperature. However, the spatial coverage is normally limited due to low density and uneven distribution of the stations. Although different interpolation techniques such as Inverse Distance Weighting (IDW), Ordinary Kriging (OK), or Multiple Linear Regression (MLR) are used to estimate air temperature from observed points, such an approach may not effectively reflect the real climatic conditions of an interpolated point. Quantifying local UHI for extensive areas based on weather stations’ observations only is not practicable. Alternatively, the use of thermal remote sensing has been widely investigated based on LST. Data from Landsat, ASTER, or MODIS have been extensively used. Indeed, LST has an indirect but significant influence on air temperatures. However, high-resolution near-surface air temperature (NSAT) is currently difficult to retrieve. Here we have experimented Geographically Weighted Regression (GWR) as an effective approach to enable NSAT estimation by accounting for spatial non-stationarity of the phenomenon. The model combines on-site measurements of air temperature, from fixed weather stations and satellite-derived LST. The approach is structured upon two main steps. First, a GWR model has been set to estimate NSAT at low resolution, by combining air temperature from discrete observations retrieved by weather stations (dependent variable) and the LST from satellite observations (predictor). At this step, MODIS data, from Terra satellite, at 1 kilometer of spatial resolution have been employed. Two time periods are considered according to satellite revisit period, i.e. 10:30 am and 9:30 pm. Afterward, the results have been downscaled at 30 meters of spatial resolution by setting a GWR model between the previously retrieved near-surface air temperature (dependent variable), the multispectral information as provided by the Landsat mission, in particular the albedo, and Digital Elevation Model (DEM) from the Shuttle Radar Topography Mission (SRTM), both at 30 meters. Albedo and DEM are now the predictors. The area under investigation is the Metropolitan City of Milan, which covers an area of approximately 1,575 km2 and encompasses a population of over 3 million inhabitants. Both models, low- (1 km) and high-resolution (30 meters), have been validated according to a cross-validation that relies on indicators such as R2, Root Mean Squared Error (RMSE) and Mean Absolute Error (MAE). All the employed indicators give evidence of highly efficient models. In addition, an alternative network of weather stations, available for the City of Milano only, has been employed for testing the accuracy of the predicted temperatures, giving and RMSE of 0.6 and 0.7 for daytime and night-time, respectively.

Keywords: urban climate, urban heat island, geographically weighted regression, remote sensing

Procedia PDF Downloads 190
6218 The Predictors of Student Engagement: Instructional Support vs Emotional Support

Authors: Tahani Salman Alangari

Abstract:

Student success can be impacted by internal factors such as their emotional well-being and external factors such as organizational support and instructional support in the classroom. This study is to identify at least one factor that forecasts student engagement. It is a cross-sectional, conducted on 6206 teachers and encompassed three years of data collection and observations of math instruction in approximately 50 schools and 300 classrooms. A multiple linear regression revealed that a model predicting student engagement from emotional support, classroom organization, and instructional support was significant. Four linear regression models were tested using hierarchical regression to examine the effects of independent variables: emotional support was the highest predictor of student engagement while instructional support was the lowest.

Keywords: student engagement, emotional support, organizational support, instructional support, well-being

Procedia PDF Downloads 74
6217 Ophthalmic Ultrasound in the Diagnosis of Retinoblastoma

Authors: Abdulrahman Algaeed

Abstract:

The Ophthalmic Ultrasound is the easiest method of early diagnosing Retinoblastoma after clinical examination. It can be done with ease without sedation. King Khaled Eye Specialist Hospital is a tertiary care center where Retinoblastoma patients are often seen and treated there. The first modality to rule out the disease is Ophthalmic Ultrasound. Classic Retinoblastoma is easily diagnosed by using the conventional 10MHz Ophthalmic Ultrasound probe in the regular clinic setup. Retinal lesion with multiple, very highly reflective surfaces within lesion typical of Calcium deposits. The use of Standardized A-scan is very useful where internal reflectivity is classified as very highly reflective. Color Doppler is extremely useful as well to show the blood flow within lesion/s. In conclusion: Ophthalmic Ultrasound should be the first tool to be used to diagnose Retinoblastoma after clinical examination. The accuracy of the Exam is very high.

Keywords: doppler, retinoblastoma, reflectivity, ultrasound

Procedia PDF Downloads 101
6216 The Solution of Nonlinear Partial Differential Equation for The Phenomenon of Instability in Homogeneous Porous Media by Homotopy Analysis Method

Authors: Kajal K. Patel, M. N. Mehta, T. R. Singh

Abstract:

When water is injected in oil formatted area in secondary oil recovery process the instability occurs near common interface due to viscosity difference of injected water and native oil. The governing equation gives rise to the non-linear partial differential equation and its solution has been obtained by Homotopy analysis method with appropriate guess value of the solution together with some conditions and standard relations. The solution gives the average cross-sectional area occupied by the schematic fingers during the occurs of instability phenomenon. The numerical and graphical presentation has developed by using Maple software.

Keywords: capillary pressure, homotopy analysis method, instability phenomenon, viscosity

Procedia PDF Downloads 491
6215 Real Estate Trend Prediction with Artificial Intelligence Techniques

Authors: Sophia Liang Zhou

Abstract:

For investors, businesses, consumers, and governments, an accurate assessment of future housing prices is crucial to critical decisions in resource allocation, policy formation, and investment strategies. Previous studies are contradictory about macroeconomic determinants of housing price and largely focused on one or two areas using point prediction. This study aims to develop data-driven models to accurately predict future housing market trends in different markets. This work studied five different metropolitan areas representing different market trends and compared three-time lagging situations: no lag, 6-month lag, and 12-month lag. Linear regression (LR), random forest (RF), and artificial neural network (ANN) were employed to model the real estate price using datasets with S&P/Case-Shiller home price index and 12 demographic and macroeconomic features, such as gross domestic product (GDP), resident population, personal income, etc. in five metropolitan areas: Boston, Dallas, New York, Chicago, and San Francisco. The data from March 2005 to December 2018 were collected from the Federal Reserve Bank, FBI, and Freddie Mac. In the original data, some factors are monthly, some quarterly, and some yearly. Thus, two methods to compensate missing values, backfill or interpolation, were compared. The models were evaluated by accuracy, mean absolute error, and root mean square error. The LR and ANN models outperformed the RF model due to RF’s inherent limitations. Both ANN and LR methods generated predictive models with high accuracy ( > 95%). It was found that personal income, GDP, population, and measures of debt consistently appeared as the most important factors. It also showed that technique to compensate missing values in the dataset and implementation of time lag can have a significant influence on the model performance and require further investigation. The best performing models varied for each area, but the backfilled 12-month lag LR models and the interpolated no lag ANN models showed the best stable performance overall, with accuracies > 95% for each city. This study reveals the influence of input variables in different markets. It also provides evidence to support future studies to identify the optimal time lag and data imputing methods for establishing accurate predictive models.

Keywords: linear regression, random forest, artificial neural network, real estate price prediction

Procedia PDF Downloads 99