Search results for: analytical reasoning
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2679

Search results for: analytical reasoning

2169 Analytic Hierarchy Process and Multi-Criteria Decision-Making Approach for Selecting the Most Effective Soil Erosion Zone in Gomati River Basin

Authors: Rajesh Chakraborty, Dibyendu Das, Rabindra Nath Barman, Uttam Kumar Mandal

Abstract:

In the present study, the objective is to find out the most effective zone causing soil erosion in the Gumati river basin located in the state of Tripura, a north eastern state of India using analytical hierarchy process (AHP) and multi-objective optimization on the basis of ratio analysis (MOORA).The watershed is segmented into 20 zones based on Area. The watershed is considered by pointing the maximum elevation from sea lever from Google earth. The soil erosion is determined using the universal soil loss equation. The different independent variables of soil loss equation bear different weightage for different soil zones. And therefore, to find the weightage factor for all the variables of soil loss equation like rainfall runoff erosivity index, soil erodibility factor etc, analytical hierarchy process (AHP) is used. And thereafter, multi-objective optimization on the basis of ratio analysis (MOORA) approach is used to select the most effective zone causing soil erosion. The MCDM technique concludes that the maximum soil erosion is occurring in the zone 14.

Keywords: soil erosion, analytic hierarchy process (AHP), multi criteria decision making (MCDM), universal soil loss equation (USLE), multi-objective optimization on the basis of ratio analysis (MOORA)

Procedia PDF Downloads 538
2168 Investigating The Effect Of Convection On The Rating Of Buried Cables Using The Finite Element Method

Authors: Sandy J. M. Balla, Jerry J. Walker, Isaac K. Kyere

Abstract:

The heat transfer coefficient at the soil–air interface is important in calculating underground cable ampacity when convection occurs. Calculating the heat transfer coefficient accurately is complex because of the temperature variations at the earth's surface. This paper presents the effect of convection heat flow across the ground surface on the rating of three single-core, 132kV, XLPE cables buried underground. The Finite element method (FEM) is a numerical analysis technique used to determine the cable rating of buried cables under installation conditions that are difficult to support when using the analytical method. This study demonstrates the use of FEM to investigate the effect of convection on the rating ofburied cables in flat formation using QuickField finite element simulation software. As a result, developing a model to simulate this type of situation necessitates important considerations such as the following boundary conditions: burial depth, soil thermal resistivity, and soil temperature, which play an important role in the simulation's accuracy and reliability. The results show that when the ground surface is taken as a convection interface, the conductor temperature rises and may exceed the maximum permissible temperature when rated current flows. This is because the ground surface acts as a convection interface between the soil and the air (fluid). This result correlates and is compared with the rating obtained using the IEC60287 analytical method, which is based on the condition that the ground surface is an isotherm.

Keywords: finite element method, convection, buried cables, steady-state rating

Procedia PDF Downloads 131
2167 Consensus Reaching Process and False Consensus Effect in a Problem of Portfolio Selection

Authors: Viviana Ventre, Giacomo Di Tollo, Roberta Martino

Abstract:

The portfolio selection problem includes the evaluation of many criteria that are difficult to compare directly and is characterized by uncertain elements. The portfolio selection problem can be modeled as a group decision problem in which several experts are invited to present their assessment. In this context, it is important to study and analyze the process of reaching a consensus among group members. Indeed, due to the various diversities among experts, reaching consensus is not necessarily always simple and easily achievable. Moreover, the concept of consensus is accompanied by the concept of false consensus, which is particularly interesting in the dynamics of group decision-making processes. False consensus can alter the evaluation and selection phase of the alternative and is the consequence of the decision maker's inability to recognize that his preferences are conditioned by subjective structures. The present work aims to investigate the dynamics of consensus attainment in a group decision problem in which equivalent portfolios are proposed. In particular, the study aims to analyze the impact of the subjective structure of the decision-maker during the evaluation and selection phase of the alternatives. Therefore, the experimental framework is divided into three phases. In the first phase, experts are sent to evaluate the characteristics of all portfolios individually, without peer comparison, arriving independently at the selection of the preferred portfolio. The experts' evaluations are used to obtain individual Analytical Hierarchical Processes that define the weight that each expert gives to all criteria with respect to the proposed alternatives. This step provides insight into how the decision maker's decision process develops, step by step, from goal analysis to alternative selection. The second phase includes the description of the decision maker's state through Markov chains. In fact, the individual weights obtained in the first phase can be reviewed and described as transition weights from one state to another. Thus, with the construction of the individual transition matrices, the possible next state of the expert is determined from the individual weights at the end of the first phase. Finally, the experts meet, and the process of reaching consensus is analyzed by considering the single individual state obtained at the previous stage and the false consensus bias. The work contributes to the study of the impact of subjective structures, quantified through the Analytical Hierarchical Process, and how they combine with the false consensus bias in group decision-making dynamics and the consensus reaching process in problems involving the selection of equivalent portfolios.

Keywords: analytical hierarchical process, consensus building, false consensus effect, markov chains, portfolio selection problem

Procedia PDF Downloads 93
2166 Medical Knowledge Management since the Integration of Heterogeneous Data until the Knowledge Exploitation in a Decision-Making System

Authors: Nadjat Zerf Boudjettou, Fahima Nader, Rachid Chalal

Abstract:

Knowledge management is to acquire and represent knowledge relevant to a domain, a task or a specific organization in order to facilitate access, reuse and evolution. This usually means building, maintaining and evolving an explicit representation of knowledge. The next step is to provide access to that knowledge, that is to say, the spread in order to enable effective use. Knowledge management in the medical field aims to improve the performance of the medical organization by allowing individuals in the care facility (doctors, nurses, paramedics, etc.) to capture, share and apply collective knowledge in order to make optimal decisions in real time. In this paper, we propose a knowledge management approach based on integration technique of heterogeneous data in the medical field by creating a data warehouse, a technique of extracting knowledge from medical data by choosing a technique of data mining, and finally an exploitation technique of that knowledge in a case-based reasoning system.

Keywords: data warehouse, data mining, knowledge discovery in database, KDD, medical knowledge management, Bayesian networks

Procedia PDF Downloads 395
2165 The Impact of Citizens’ Involvement on Their Perception of the Brand’s Image: The Case of the City of Casablanca

Authors: Abderrahmane Mousstain, Ez-Zohra Belkadi

Abstract:

Many authors support more participatory and inclusive place branding practices that empower stakeholders’ participation. According to this participatory point of view, the effectiveness of place branding strategies cannot be achieved without citizen involvement. However, the role of all residents as key participants in the city branding process has not been widely discussed. The aim of this paper was to determine how citizens’ involvement impacts their perceptions of the city's image, using a multivariate model. To test our hypotheses hypothetical-deductive reasoning by the quantitative method was chosen. Our investigation is based on data collected through a survey among 200 citizens of Casablanca. Results show that the more citizens are involved, the more they tend to evaluate the image of the brand positively. Additionally, the degree of involvement seems to impact satisfaction and a sense of belonging. As well, the more citizen develops a sense of belonging to the city, the more favorable his or her perception of the brand image is. Ultimately, the role of citizens shouldn’t be limited to reception. They are also Co-creators of the brand, who ensure the correlation of the brand with authentic place roots.

Keywords: citybranding, sense of belonging, satisfaction, impact, brand’s image

Procedia PDF Downloads 176
2164 Occurrence of Foreign Matter in Food: Applied Identification Method - Association of Official Agricultural Chemists (AOAC) and Food and Drug Administration (FDA)

Authors: E. C. Mattos, V. S. M. G. Daros, R. Dal Col, A. L. Nascimento

Abstract:

The aim of this study is to present the results of a retrospective survey on the foreign matter found in foods analyzed at the Adolfo Lutz Institute, from July 2001 to July 2015. All the analyses were conducted according to the official methods described on Association of Official Agricultural Chemists (AOAC) for the micro analytical procedures and Food and Drug Administration (FDA) for the macro analytical procedures. The results showed flours, cereals and derivatives such as baking and pasta products were the types of food where foreign matters were found more frequently followed by condiments and teas. Fragments of stored grains insects, its larvae, nets, excrement, dead mites and rodent excrement were the most foreign matter found in food. Besides, foreign matters that can cause a physical risk to the consumer’s health such as metal, stones, glass, wood were found but rarely. Miscellaneous (shell, sand, dirt and seeds) were also reported. There are a lot of extraneous materials that are considered unavoidable since are something inherent to the product itself, such as insect fragments in grains. In contrast, there are avoidable extraneous materials that are less tolerated because it is preventable with the Good Manufacturing Practice. The conclusion of this work is that although most extraneous materials found in food are considered unavoidable it is necessary to keep the Good Manufacturing Practice throughout the food processing as well as maintaining a constant surveillance of the production process in order to avoid accidents that may lead to occurrence of these extraneous materials in food.

Keywords: extraneous materials, food contamination, foreign matter, surveillance

Procedia PDF Downloads 359
2163 Accuracy of VCCT for Calculating Stress Intensity Factor in Metal Specimens Subjected to Bending Load

Authors: Sanjin Kršćanski, Josip Brnić

Abstract:

Virtual Crack Closure Technique (VCCT) is a method used for calculating stress intensity factor (SIF) of a cracked body that is easily implemented on top of basic finite element (FE) codes and as such can be applied on the various component geometries. It is a relatively simple method that does not require any special finite elements to be used and is usually used for calculating stress intensity factors at the crack tip for components made of brittle materials. This paper studies applicability and accuracy of VCCT applied on standard metal specimens containing trough thickness crack, subjected to an in-plane bending load. Finite element analyses were performed using regular 4-node, regular 8-node and a modified quarter-point 8-node 2D elements. Stress intensity factor was calculated from the FE model results for a given crack length, using data available from FE analysis and a custom programmed algorithm based on virtual crack closure technique. Influence of the finite element size on the accuracy of calculated SIF was also studied. The final part of this paper includes a comparison of calculated stress intensity factors with results obtained from analytical expressions found in available literature and in ASTM standard. Results calculated by this algorithm based on VCCT were found to be in good correlation with results obtained with mentioned analytical expressions.

Keywords: VCCT, stress intensity factor, finite element analysis, 2D finite elements, bending

Procedia PDF Downloads 305
2162 Capturing Healthcare Expert’s Knowledge Digitally: A Scoping Review of Current Approaches

Authors: Sinead Impey, Gaye Stephens, Declan O’Sullivan

Abstract:

Mitigating organisational knowledge loss presents challenges for knowledge managers. Expert knowledge is embodied in people and captured in ‘routines, processes, practices and norms’ as well as in the paper system. These knowledge stores have limitations in so far as they make knowledge diffusion beyond geography or over time difficult. However, technology could present a potential solution by facilitating the capture and management of expert knowledge in a codified and sharable format. Before it can be digitised, however, the knowledge of healthcare experts must be captured. Methods: As a first step in a larger project on this topic, a scoping review was conducted to identify how expert healthcare knowledge is captured digitally. The aim of the review was to identify current healthcare knowledge capture practices, identify gaps in the literature, and justify future research. The review followed a scoping review framework. From an initial 3,430 papers retrieved, 22 were deemed relevant and included in the review. Findings: Two broad approaches –direct and indirect- with themes and subthemes emerged. ‘Direct’ describes a process whereby knowledge is taken directly from subject experts. The themes identified were: ‘Researcher mediated capture’ and ‘Digital mediated capture’. The latter was further distilled into two sub-themes: ‘Captured in specified purpose platforms (SPP)’ and ‘Captured in a virtual community of practice (vCoP)’. ‘Indirect’ processes rely on extracting new knowledge using artificial intelligence techniques from previously captured data. Using this approach, the theme ‘Generated using artificial intelligence methods’ was identified. Although presented as distinct themes, some papers retrieved discuss combining more than one approach to capture knowledge. While no approach emerged as superior, two points arose from the literature. Firstly, human input was evident across themes, even with indirect approaches. Secondly, a range of challenges common among approaches was highlighted. These were (i) ‘Capturing an expert’s knowledge’- Difficulties surrounding capturing an expert’s knowledge related to identifying the ‘expert’ say from the very experienced and how to capture their tacit or difficult to articulate knowledge. (ii) ‘Confirming quality of knowledge’- Once captured, challenges noted surrounded how to validate knowledge captured and, therefore, quality. (iii) ‘Continual knowledge capture’- Once knowledge is captured, validated, and used in a system; however, the process is not complete. Healthcare is a knowledge-rich environment with new evidence emerging frequently. As such, knowledge needs to be reviewed, updated, or removed (redundancy) as appropriate. Although some methods were proposed to address this, such as plausible reasoning or case-based reasoning, conclusions could not be drawn from the papers retrieved. It was, therefore, highlighted as an area for future research. Conclusion: The results described two broad approaches – direct and indirect. Three themes were identified: ‘Researcher mediated capture (Direct)’; ‘Digital mediated capture (Direct)’ and ‘Generated using artificial intelligence methods (Indirect)’. While no single approach was deemed superior, common challenges noted among approaches were: ‘capturing an expert’s knowledge’, ‘confirming quality of knowledge’, and ‘continual knowledge capture’. However, continual knowledge capture was not fully explored in the papers retrieved and was highlighted as an important area for future research. Acknowledgments: This research is partially funded by the ADAPT Centre under the SFI Research Centres Programme (Grant 13/RC/2106) and is co-funded under the European Regional Development Fund.

Keywords: expert knowledge, healthcare, knowledge capture and knowledge management

Procedia PDF Downloads 134
2161 Simplified Stress Gradient Method for Stress-Intensity Factor Determination

Authors: Jeries J. Abou-Hanna

Abstract:

Several techniques exist for determining stress-intensity factors in linear elastic fracture mechanics analysis. These techniques are based on analytical, numerical, and empirical approaches that have been well documented in literature and engineering handbooks. However, not all techniques share the same merit. In addition to overly-conservative results, the numerical methods that require extensive computational effort, and those requiring copious user parameters hinder practicing engineers from efficiently evaluating stress-intensity factors. This paper investigates the prospects of reducing the complexity and required variables to determine stress-intensity factors through the utilization of the stress gradient and a weighting function. The heart of this work resides in the understanding that fracture emanating from stress concentration locations cannot be explained by a single maximum stress value approach, but requires use of a critical volume in which the crack exists. In order to understand the effectiveness of this technique, this study investigated components of different notch geometry and varying levels of stress gradients. Two forms of weighting functions were employed to determine stress-intensity factors and results were compared to analytical exact methods. The results indicated that the “exponential” weighting function was superior to the “absolute” weighting function. An error band +/- 10% was met for cases ranging from a steep stress gradient in a sharp v-notch to the less severe stress transitions of a large circular notch. The incorporation of the proposed method has shown to be a worthwhile consideration.

Keywords: fracture mechanics, finite element method, stress intensity factor, stress gradient

Procedia PDF Downloads 135
2160 Towards Efficient Reasoning about Families of Class Diagrams Using Union Models

Authors: Tejush Badal, Sanaa Alwidian

Abstract:

Class diagrams are useful tools within the Unified Modelling Language (UML) to model and visualize the relationships between, and properties of objects within a system. As a system evolves over time and space (e.g., products), a series of models with several commonalities and variabilities create what is known as a model family. In circumstances where there are several versions of a model, examining each model individually, becomes expensive in terms of computation resources. To avoid performing redundant operations, this paper proposes an approach for representing a family of class diagrams into Union Models to represent model families using a single generic model. The paper aims to analyze and reason about a family of class diagrams using union models as opposed to individual analysis of each member model in the family. The union algorithm provides a holistic view of the model family, where the latter cannot be otherwise obtained from an individual analysis approach, this in turn, enhances the analysis performed in terms of speeding up the time needed to analyze a family of models together as opposed to analyzing individual models, one model at a time.

Keywords: analysis, class diagram, model family, unified modeling language, union model

Procedia PDF Downloads 74
2159 Analytical Study and Conservation Processes of a Wooden Coffin of Middel Kingdom, Ancient Egypt

Authors: Mohamed Ahmed Abd El Kader

Abstract:

This paper describes the conservation processes of an Ancient Egyptian wooden coffin dating back to the Middle Kingdom, ancient Egypt, using several scientific and analytical methods in order to provide a deeper understanding of the deterioration status and a greater awareness of how well preserved the object is. Visual observation and 2D Programs, as well as Optical Microscopy (OM), Environmental scanning Electron Microscopy (ESEM), X-ray Diffraction (XRD) and Fourier Transform Infrared Spectroscopy (FTIR) were used in our study. The identification of wood species and the composition of the pigments and previous restoration materials were made. The coffin was previously conserved and stored in improper conditions, which led to its further deterioration; the surface of the lid dust, which obscured the decorations as well as all necessary restoration work was promptly carried out as soon as the coffin was transferred from the display hall from the Egyptian Museum to the Wood Conservation Laboratory of the Grand Egyptian Museum-Conservation Center (GEM-CC). The analyses provided detailed information concerning the original materials and the materials added during the previous treatment interventions, which was considered when applying the conservation plan. Conservation procedures have been applied with high accuracy to conserve the coffin including cleaning, consolidation of fragile painted layers, and the wooden boards forming the sides of the coffin were reassembled in their original positions. The materials and methods that were applied were extremely effective in stability and reinforcement of the coffin without harmfulness to the original materials and the coffin was successfully conserved and ready to display in the Grand Egyptian Museum (GEM).

Keywords: coffin, middle kingdom, deterioration, 2d program

Procedia PDF Downloads 53
2158 Childhood Trauma and Borderline Personality: An Analysis of the Root Causes and Treatment Plans

Authors: Sidika McNeil

Abstract:

Borderline personality disorder (BPD) is a personality disorder that has been found to have strong origins in childhood trauma. One of the key symptoms of BPD is an association with irregular moods swings, as well as suicidal ideation (SI). Owing to the typically severe trauma patients experience during childhood, it is hard for them to control their emotions and thus makes it hard to emotionally regulate. It is then very common for those suffering from BPD to turn to unhealthy coping mechanisms, such as substance use, unhealthy relationships, and more, often unsuccessfully creating experiences that facilitate safety which leads to further negative experiences. With the high suicide rating among children, adolescents, and teens, and an ever-increasing number of children being diagnosed with BPD, it is very important that more research is done to find further treatments for patients who are currently suffering. Methods: Utilizing data found in prior studies, this paper will analyze the literature to focus on a comprehensive treatment plan for those with DBT. It is currently suggested that with the use of dialectical behavioral therapy (DBT), a therapy that focuses on changing negative thinking patterns and pushes for more positive ones is helpful for treatment for those with BPD. Though this therapy is not a cure to BPD, it does help mitigate the risk; this essay will explore other options that can further the treatment process, such as cognitive analytical therapy (CAT), which focuses on delving into the past to find the root causes of an issue to create coping strategies and harm reduction, a type of therapy used to aid patients in lowering the use of substances without complete cessation. Results: The research provides enough evidence to link between the treatment of BPD with the utilization of CAT.

Keywords: borderline personality disorder, cognitive analytical therapy, dialectical behavioral therapy, harm reduction, suicidal ideation

Procedia PDF Downloads 176
2157 Analytical Validity Of A Tech Transfer Solution To Internalize Genetic Testing

Authors: Lesley Northrop, Justin DeGrazia, Jessica Greenwood

Abstract:

ASPIRA Labs now offers an en-suit and ready-to-implement technology transfer solution to enable labs and hospitals that lack the resources to build it themselves to offer in-house genetic testing. This unique platform employs a patented Molecular Inversion Probe (MIP) technology that combines the specificity of a hybrid capture protocol with the ease of an amplicon-based protocol and utilizes an advanced bioinformatics analysis pipeline based on machine learning. To demonstrate its efficacy, two independent genetic tests were validated on this technology transfer platform: expanded carrier screening (ECS) and hereditary cancer testing (HC). The analytical performance of ECS and HC was validated separately in a blinded manner for calling three different types of variants: SNVs, short indels (typically, <50 bp), and large indels/CNVs defined as multi-exonic del/dup events. The reference set was constructed using samples from Coriell Institute, an external clinical genetic testing laboratory, Maine Molecular Quality Controls Inc. (MMQCI), SeraCare and GIAB Consortium. Overall, the analytical performance showed a sensitivity and specificity of >99.4% for both ECS and HC in detecting SNVs. For indels, both tests reported specificity of 100%, and ECS demonstrated a sensitivity of 100%, whereas HC exhibited a sensitivity of 96.5%. The bioinformatics pipeline also correctly called all reference CNV events resulting in a sensitivity of 100% for both tests. No additional calls were made in the HC panel, leading to a perfect performance (specificity and F-measure of 100%). In the carrier panel, however, three additional positive calls were made outside the reference set. Two of these calls were confirmed using an orthogonal method and were re-classified as true positives leaving only one false positive. The pipeline also correctly identified all challenging carrier statuses, such as positive cases for spinal muscular atrophy and alpha-thalassemia, resulting in 100% sensitivity. After confirmation of additional positive calls via long-range PCR and MLPA, specificity for such cases was estimated at 99%. These performance metrics demonstrate that this tech-transfer solution can be confidently internalized by clinical labs and hospitals to offer mainstream ECS and HC as part of their test catalog, substantially increasing access to quality germline genetic testing for labs of all sizes and resources levels.

Keywords: clinical genetics, genetic testing, molecular genetics, technology transfer

Procedia PDF Downloads 178
2156 The Game of Dominoes as Teaching-Learning Method of Basic Concepts of Differential Calculus

Authors: Luis Miguel Méndez Díaz

Abstract:

In this article, a mathematics teaching-learning strategy will be presented, specifically differential calculus in one variable, in a fun and competitive space in which the action on the part of the student is manifested and not only the repetition of information on the part of the teacher. Said action refers to motivating, problematizing, summarizing, and coordinating a game of dominoes whose thematic cards are designed around the basic and main contents of differential calculus. The strategies for teaching this area are diverse and precisely the game of dominoes is one of the most used strategies in the practice of mathematics because it stimulates logical reasoning and mental abilities. The objective on this investigation is to identify the way in which the game of dominoes affects the learning and understanding of fundamentals concepts of differential calculus in one variable through experimentation carried out on students of the first semester of the School of Engineering and Sciences of the Technological Institute of Monterrey Campus Querétaro. Finally, the results of this study will be presented and the use of this strategy in other topics around mathematics will be recommended to facilitate logical and meaningful learning in students.

Keywords: collaborative learning, logical-mathematical intelligence, mathematical games, multiple intelligences

Procedia PDF Downloads 84
2155 An Analytical Approach of Computational Complexity for the Method of Multifluid Modelling

Authors: A. K. Borah, A. K. Singh

Abstract:

In this paper we deal building blocks of the computer simulation of the multiphase flows. Whole simulation procedure can be viewed as two super procedures; The implementation of VOF method and the solution of Navier Stoke’s Equation. Moreover, a sequential code for a Navier Stoke’s solver has been studied.

Keywords: Bi-conjugate gradient stabilized (Bi-CGSTAB), ILUT function, krylov subspace, multifluid flows preconditioner, simple algorithm

Procedia PDF Downloads 528
2154 Accurate Binding Energy of Ytterbium Dimer from Ab Initio Calculations and Ultracold Photoassociation Spectroscopy

Authors: Giorgio Visentin, Alexei A. Buchachenko

Abstract:

Recent proposals to use Yb dimer as an optical clock and as a sensor for non-Newtonian gravity imply the knowledge of its interaction potential. Here, the ground-state Born-Oppenheimer Yb₂ potential energy curve is represented by a semi-analytical function, consisting of short- and long-range contributions. For the former, the systematic ab initio all-electron exact 2-component scalar-relativistic CCSD(T) calculations are carried out. Special care is taken to saturate diffuse basis set component with the atom- and bond-centered primitives and reach the complete basis set limit through n = D, T, Q sequence of the correlation-consistent polarized n-zeta basis sets. Similar approaches are used to the long-range dipole and quadrupole dispersion terms by implementing the CCSD(3) polarization propagator method for dynamic polarizabilities. Dispersion coefficients are then computed through Casimir-Polder integration. The semiclassical constraint on the number of the bound vibrational levels known for the ¹⁷⁴Yb isotope is used to scale the potential function. The scaling, based on the most accurate ab initio results, bounds the interaction energy of two Yb atoms within the narrow 734 ± 4 cm⁻¹ range, in reasonable agreement with the previous ab initio-based estimations. The resulting potentials can be used as the reference for more sophisticated models that go beyond the Born-Oppenheimer approximation and provide the means of their uncertainty estimations. The work is supported by Russian Science Foundation grant # 17-13-01466.

Keywords: ab initio coupled cluster methods, interaction potential, semi-analytical function, ytterbium dimer

Procedia PDF Downloads 154
2153 An Analysis of Human Resource Management Policies for Constructing Employer Brands in the Logistics Sector

Authors: Müberra Yüksel, Ömer Faruk Görçün

Abstract:

The purpose of the present study is to investigate the role of strategic human resource management (SHRM) in constructing "employer branding" in logistics. Prior research does not focus on internal stakeholders, that is, employees. Despite the fact that logistic sector has become customer-oriented, the focus is solely on service quality as the unique aspect of logistic companies for competitive advantage. With an increasing interest lately in internal marketing of the employer brand, the emphasis is on the value that human capital brings to the firm which cannot be imitated. `Employer branding` has been the application of branding and relationship marketing principles for competitive advantage in SHRM. Employer branding is an organizing framework for human resource managers since it represents an organization’s efforts to promote, both within and outside, a coherent view of what makes the firm different and desirable as an employer, i.e., the distinct “employer brand personality” and "employee value propositions" (EVP) offered. The presumption of employer branding enhanced by internal marketing is to make customer-conscious employees to handle services better by being aligned with business mission and goals. Starting from internal customers and analyzing the gaps of EVP by using analytical hierarchy process methodology (AHP) and inquiring whether these brand values are communicated and conceived well may be the initial steps in our proposal for employer branding in logistics sector. This empirical study aims to fill this research gap within the context of an emergent market- Turkey, which is located at a hub of transportation and logistics.

Keywords: Strategic Human Resource Management (SHRM), employer branding, Employee Value Propositions (EVP), Analytical Hierarchy Process (AHP), logistics

Procedia PDF Downloads 343
2152 Study of Human Position in Architecture with Contextual Approach

Authors: E. Zarei, M. Bazaei, A. seifi, A. Keshavarzi

Abstract:

Contextuallism has been always the main component of urban science. It not only has great direct and indirect impact on behaviors, events and interactions, but also is one of the basic factors of an urban values and identity. Nowadays there might be some deficiencies in the cities. In the theories of environment designing, humanistic orientations with the focus on culture and cultural variables would enable us to transfer information. To communicate with the context in which human lives, he needs some common memories, understandable symbols and daily activities in that context. The configuration of a place can impact on human’s behaviors. The goal of this research is to review 7 projects in different parts of the world with various usages and some factors such as ‘sense of place’, ‘sense of belonging’ and ‘social and cultural relations’ will be discussed in these projects. The method used for research in this project is descriptive- analytic. Library information and Internet are the main sources of gathering information and the method of reasoning used in this project is inductive. The consequence of this research will be some data in the form of tables that has been extracted from mentioned projects.

Keywords: contextuallism with humanistic approach, sense of place, sense of belonging, social and cultural relations

Procedia PDF Downloads 396
2151 Bioinformatic Approaches in Population Genetics and Phylogenetic Studies

Authors: Masoud Sheidai

Abstract:

Biologists with a special field of population genetics and phylogeny have different research tasks such as populations’ genetic variability and divergence, species relatedness, the evolution of genetic and morphological characters, and identification of DNA SNPs with adaptive potential. To tackle these problems and reach a concise conclusion, they must use the proper and efficient statistical and bioinformatic methods as well as suitable genetic and morphological characteristics. In recent years application of different bioinformatic and statistical methods, which are based on various well-documented assumptions, are the proper analytical tools in the hands of researchers. The species delineation is usually carried out with the use of different clustering methods like K-means clustering based on proper distance measures according to the studied features of organisms. A well-defined species are assumed to be separated from the other taxa by molecular barcodes. The species relationships are studied by using molecular markers, which are analyzed by different analytical methods like multidimensional scaling (MDS) and principal coordinate analysis (PCoA). The species population structuring and genetic divergence are usually investigated by PCoA and PCA methods and a network diagram. These are based on bootstrapping of data. The Association of different genes and DNA sequences to ecological and geographical variables is determined by LFMM (Latent factor mixed model) and redundancy analysis (RDA), which are based on Bayesian and distance methods. Molecular and morphological differentiating characters in the studied species may be identified by linear discriminant analysis (DA) and discriminant analysis of principal components (DAPC). We shall illustrate these methods and related conclusions by giving examples from different edible and medicinal plant species.

Keywords: GWAS analysis, K-Means clustering, LFMM, multidimensional scaling, redundancy analysis

Procedia PDF Downloads 124
2150 Modeling Depth Averaged Velocity and Boundary Shear Stress Distributions

Authors: Ebissa Gadissa Kedir, C. S. P. Ojha, K. S. Hari Prasad

Abstract:

In the present study, the depth-averaged velocity and boundary shear stress in non-prismatic compound channels with three different converging floodplain angles ranging from 1.43ᶱ to 7.59ᶱ have been studied. The analytical solutions were derived by considering acting forces on the channel beds and walls. In the present study, five key parameters, i.e., non-dimensional coefficient, secondary flow term, secondary flow coefficient, friction factor, and dimensionless eddy viscosity, were considered and discussed. An expression for non-dimensional coefficient and integration constants was derived based on the boundary conditions. The model was applied to different data sets of the present experiments and experiments from other sources, respectively, to examine and analyse the influence of floodplain converging angles on depth-averaged velocity and boundary shear stress distributions. The results show that the non-dimensional parameter plays important in portraying the variation of depth-averaged velocity and boundary shear stress distributions with different floodplain converging angles. Thus, the variation of the non-dimensional coefficient needs attention since it affects the secondary flow term and secondary flow coefficient in both the main channel and floodplains. The analysis shows that the depth-averaged velocities are sensitive to a shear stress-dependent model parameter non-dimensional coefficient, and the analytical solutions are well agreed with experimental data when five parameters are included. It is inferred that the developed model may facilitate the interest of others in complex flow modeling.

Keywords: depth-average velocity, converging floodplain angles, non-dimensional coefficient, non-prismatic compound channels

Procedia PDF Downloads 74
2149 Empowering the Sustainability of Community Health: An Application of the Theory of Maqasid Al-Shariah

Authors: Ahasanul Haque, Noor Hazilah Abd Manaf, Zohurul Anis, Tarekol Islam

Abstract:

Sustainable community health (SCH) is an example of a new healthcare concept formed from applying the Maqasid al-Shariah principle to hospital management and delivery services. Because the idea is novel, it needs comprehensive and ongoing investigation to be improved. However, there is a lack of research on the necessity of developing sustainable community health (SCH), particularly its organizational structure. Furthermore, there is a misconception about the order of components in Maqasid al-Shariah, particularly in a hospital setting. Furthermore, the use of medicines and treatment by conventional recommendations to carry out the treatment by the Maqasid al Shariah. As such, this study focuses on the essential prerequisite for establishing a sustainable community health system based on Maqasid al-Shariah. This study discusses the use of Maqasid al-Shariah in administration and treatment. In this qualitative research approach, a literature search and interviews with specialists are conducted. The gathered data is examined using content analysis, emphasizing inductive and deductive reasoning. The research reveals that the Shariah Advisory Council and Shariah Critical Point are necessary for sustainable community health. In conclusion, by discussing the causes for each instance, this research adds to the creation of methods for determining the level of Maasid al-Shariah in-hospital care.

Keywords: empowering, sustainability, community health, maqasid al shariah, hospital and malaysia

Procedia PDF Downloads 84
2148 Exploring Antifragility Principles in Humanitarian Supply Chain: The key Role of Information Systems

Authors: Sylvie Michel, Sylvie Gerbaix, Marc Bidan

Abstract:

The COVID-19 pandemic has been a major and global disruption that has affected all supply chains on a worldwide scale. Consequently, the question posed by this communication is to understand how - in the face of such disruptions - supply chains, including their actors, management tools, and processes, react, survive, adapt, and even improve. To do so, the concepts of resilience and antifragility applied to a supply chain have been leveraged. This article proposes to perceive resilience as a step to surpass in moving towards antifragility. The research objective is to propose an analytical framework to measure and compare resilience and antifragility, with antifragility seen as a property of a system that improves when subjected to disruptions rather than merely resisting these disruptions, as is the case with resilience. A unique case study was studied - MSF logistics (France) - using a qualitative methodology. Semi-structured interviews were conducted in person and remotely in multiple phases: during and immediately after the COVID crisis (8 interviews from March 2020 to April 2021), followed by a new round from September to November 2023. A Delphi method was employed. The interviews were analyzed using coding and a thematic framework. One of the theoretical contributions is consolidating the field of supply chain resilience research by precisely characterizing the dimensions of resilience for a humanitarian supply chain (Reorganization, Collaboration mediated by IS, Humanitarian culture). In this regard, a managerial contribution of this study is providing a guide for managers to identify the four dimensions and sub-dimensions of supply chain resilience. This enables managers to focus their decisions and actions on dimensions that will enhance resilience. Most importantly, another contribution is comparing the concepts of resilience and antifragility and proposing an analytical framework for antifragility—namely, the mechanisms on which MSF logistics relied to capitalize on uncertainties, contingencies, and shocks rather than simply enduring them. For MSF Logistics, antifragility manifested through the ability to identify opportunities hidden behind the uncertainties and shocks of COVID-19, reducing vulnerability, and fostering a culture that encourages innovation and the testing of new ideas. Logistics, particularly in the humanitarian domain, must be able to adapt to environmental disruptions. In this sense, this study identifies and characterizes the dimensions of resilience implemented by humanitarian logistics. Moreover, this research goes beyond the concept of resilience to propose an analytical framework for the concept of antifragility. The organization studied emerged stronger from the COVID-19 crisis due to the mechanisms we identified, allowing us to characterize antifragility. Finally, the results show that the information system plays a key role in antifragility.

Keywords: antifragility, humanitarian supply chain, information systems, qualitative research, resilience.

Procedia PDF Downloads 64
2147 The Use of Language as a Cognitive Tool in French Immersion Teaching

Authors: Marie-Josée Morneau

Abstract:

A literacy-based approach, centred on the use of the language of instruction as a cognitive tool, can increase the L2 communication skills of French immersion students. Academic subject areas such as science and mathematics offer an authentic language learning context where students can become more proficient speakers while using specific vocabulary and language structures to learn, interact and communicate their reasoning, when provided the opportunities and guidance to do so. In this Canadian quasi-experimental study, the effects of teaching specific language elements during mathematic classes through literacy-based activities in Early French Immersion programming were compared between two Grade 7/8 groups: the experimental group, which received literacy-based teaching for a 6-week period, and the control group, which received regular teaching instruction. The results showed that the participants from the experimental group made more progress in their mathematical communication skills, which suggests that targeting L2 language as a cognitive tool can be beneficial to immersion learners who learn mathematic concepts and remind us that all L2 teachers are language teachers.

Keywords: mathematics, French immersion, literacy-based, oral communication, L2

Procedia PDF Downloads 76
2146 Replicating Brain’s Resting State Functional Connectivity Network Using a Multi-Factor Hub-Based Model

Authors: B. L. Ho, L. Shi, D. F. Wang, V. C. T. Mok

Abstract:

The brain’s functional connectivity while temporally non-stationary does express consistency at a macro spatial level. The study of stable resting state connectivity patterns hence provides opportunities for identification of diseases if such stability is severely perturbed. A mathematical model replicating the brain’s spatial connections will be useful for understanding brain’s representative geometry and complements the empirical model where it falls short. Empirical computations tend to involve large matrices and become infeasible with fine parcellation. However, the proposed analytical model has no such computational problems. To improve replicability, 92 subject data are obtained from two open sources. The proposed methodology, inspired by financial theory, uses multivariate regression to find relationships of every cortical region of interest (ROI) with some pre-identified hubs. These hubs acted as representatives for the entire cortical surface. A variance-covariance framework of all ROIs is then built based on these relationships to link up all the ROIs. The result is a high level of match between model and empirical correlations in the range of 0.59 to 0.66 after adjusting for sample size; an increase of almost forty percent. More significantly, the model framework provides an intuitive way to delineate between systemic drivers and idiosyncratic noise while reducing dimensions by more than 30 folds, hence, providing a way to conduct attribution analysis. Due to its analytical nature and simple structure, the model is useful as a standalone toolkit for network dependency analysis or as a module for other mathematical models.

Keywords: functional magnetic resonance imaging, multivariate regression, network hubs, resting state functional connectivity

Procedia PDF Downloads 151
2145 Fin Efficiency of Helical Fin with Fixed Fin Tip Temperature Boundary Condition

Authors: Richard G. Carranza, Juan Ospina

Abstract:

The fin efficiency for a helical fin with a fixed fin tip (or arbitrary) temperature boundary condition is presented. Firstly, the temperature profile throughout the fin is determined via an energy balance around the fin itself. Secondly, the fin efficiency is formulated by integrating across the entire surface of the helical fin. An analytical expression for the fin efficiency is presented and compared with the literature for accuracy.

Keywords: efficiency, fin, heat, helical, transfer

Procedia PDF Downloads 684
2144 Detection and Quantification of Ochratoxin A in Food by Aptasensor

Authors: Moez Elsaadani, Noel Durand, Brice Sorli, Didier Montet

Abstract:

Governments and international instances are trying to improve the food safety system to prevent, reduce or avoid the increase of food borne diseases. This food risk is one of the major concerns for the humanity. The contamination by mycotoxins is a threat to the health and life of humans and animals. One of the most common mycotoxin contaminating feed and foodstuffs is Ochratoxin A (OTA), which is a secondary metabolite, produced by Aspergillus and Penicillium strains. OTA has a chronic toxic effect and proved to be mutagenic, nephrotoxic, teratogenic, immunosuppressive, and carcinogenic. On the other side, because of their high stability, specificity, affinity, and their easy chemical synthesis, aptamer based methods are applied to OTA biosensing as alternative to traditional analytical technique. In this work, five aptamers have been tested to confirm qualitatively and quantitatively their binding with OTA. In the same time, three different analytical methods were tested and compared based on their ability to detect and quantify the OTA. The best protocol that was established to quantify free OTA from linked OTA involved an ultrafiltration method in green coffee solution with. OTA was quantified by HPLC-FLD to calculate the binding percentage of all five aptamers. One aptamer (The most effective with 87% binding with OTA) has been selected to be our biorecognition element to study its electrical response (variation of electrical properties) in the presence of OTA in order to be able to make a pairing with a radio frequency identification (RFID). This device, which is characterized by its low cost, speed, and a simple wireless information transmission, will implement the knowledge on the mycotoxins molecular sensors (aptamers), an electronic device that will link the information, the quantification and make it available to operators.

Keywords: aptamer, aptasensor, detection, Ochratoxin A

Procedia PDF Downloads 181
2143 Composite Coatings of Piezoelectric Quartz Sensors Based on Viscous Sorbents and Casein Micelles

Authors: Shuba Anastasiia, Kuchmenko Tatiana, Umarkhanov Ruslan

Abstract:

The development of new sensitive coatings for sensors is one of the key directions in the development of sensor technologies. Recently, there has been a trend towards the creation of multicomponent coatings for sensors, which make it possible to increase the sensitivity, and specificity, and improve the performance properties of sensors. When analyzing samples with a complex matrix of biological origin, the inclusion of micelles of bioactive substances (amino and nucleic acids, peptides, proteins) in the composition of the sensor coating can also increase useful analytical information. The purpose of this work is to evaluate the analytical characteristics of composite coatings of piezoelectric quartz sensors based on medium-molecular viscous sorbents with incorporated micellar casein concentrate during the sorption of vapors of volatile organic compounds. The sorption properties of the coatings were studied by piezoelectric quartz microbalance. Macromolecular compounds (dicyclohexyl-18-crown-6, triton X-100, lanolin, micellar casein concentrate) were used as sorbents. Highly volatile organic compounds of various classes (alcohols, acids, aldehydes, esters) and water were selected as test substances. It has been established that composite coatings of sensors with the inclusion of micellar casein are more stable and selective to vapors of highly volatile compounds than to water vapors. The method and technique of forming a composite coating using molecular viscous sorbents do not affect the kinetic features of VOC sorption. When casein micelles are used, the features of kinetic sorption depend on the matrix of the coating.

Keywords: piezoquartz sensor, viscous sorbents, micellar casein, coating, volatile compounds

Procedia PDF Downloads 122
2142 Representational Issues in Learning Solution Chemistry at Secondary School

Authors: Lam Pham, Peter Hubber, Russell Tytler

Abstract:

Students’ conceptual understandings of chemistry concepts/phenomena involve capability to coordinate across the three levels of Johnston’s triangle model. This triplet model is based on reasoning about chemical phenomena across macro, sub-micro and symbolic levels. In chemistry education, there is a need for further examining inquiry-based approaches that enhance students’ conceptual learning and problem solving skills. This research adopted a directed inquiry pedagogy based on students constructing and coordinating representations, to investigate senior school students’ capabilities to flexibly move across Johnston’ levels when learning dilution and molar concentration concepts. The participants comprise 50 grade 11 and 20 grade 10 students and 4 chemistry teachers who were selected from 4 secondary schools located in metropolitan Melbourne, Victoria. This research into classroom practices used ethnographic methodology, involved teachers working collaboratively with the research team to develop representational activities and lesson sequences in the instruction of a unit on solution chemistry. The representational activities included challenges (Representational Challenges-RCs) that used ‘representational tools’ to assist students to move across Johnson’s three levels for dilution phenomena. In this report, the ‘representational tool’ called ‘cross and portion’ model was developed and used in teaching and learning the molar concentration concept. Students’ conceptual understanding and problem solving skills when learning with this model are analysed through group case studies of year 10 and 11 chemistry students. In learning dilution concepts, students in both group case studies actively conducted a practical experiment, used their own language and visualisation skills to represent dilution phenomena at macroscopic level (RC1). At the sub-microscopic level, students generated and negotiated representations of the chemical interactions between solute and solvent underpinning the dilution process. At the symbolic level, students demonstrated their understandings about dilution concepts by drawing chemical structures and performing mathematical calculations. When learning molar concentration with a ‘cross and portion’ model (RC2), students coordinated across visual and symbolic representational forms and Johnson’s levels to construct representations. The analysis showed that in RC1, Year 10 students needed more ‘scaffolding’ in inducing to representations to explicit the form and function of sub-microscopic representations. In RC2, Year 11 students showed clarity in using visual representations (drawings) to link to mathematics to solve representational challenges about molar concentration. In contrast, year 10 students struggled to get match up the two systems, symbolic system of mole per litre (‘cross and portion’) and visual representation (drawing). These conceptual problems do not lie in the students’ mathematical calculation capability but rather in students’ capability to align visual representations with the symbolic mathematical formulations. This research also found that students in both group case studies were able to coordinate representations when probed about the use of ‘cross and portion’ model (in RC2) to demonstrate molar concentration of diluted solutions (in RC1). Students mostly succeeded in constructing ‘cross and portion’ models to represent the reduction of molar concentration of the concentration gradients. In conclusion, this research demonstrated how the strategic introduction and coordination of chemical representations across modes and across the macro, sub-micro and symbolic levels, supported student reasoning and problem solving in chemistry.

Keywords: cross and portion, dilution, Johnston's triangle, molar concentration, representations

Procedia PDF Downloads 137
2141 Ontology for Cross-Site-Scripting (XSS) Attack in Cybersecurity

Authors: Jean Rosemond Dora, Karol Nemoga

Abstract:

In this work, we tackle a frequent problem that frequently occurs in the cybersecurity field which is the exploitation of websites by XSS attacks, which are nowadays considered a complicated attack. These types of attacks aim to execute malicious scripts in a web browser of the client by including code in a legitimate web page. A serious matter is when a website accepts the “user-input” option. Attackers can exploit the web application (if vulnerable), and then steal sensitive data (session cookies, passwords, credit cards, etc.) from the server and/or from the client. However, the difficulty of the exploitation varies from website to website. Our focus is on the usage of ontology in cybersecurity against XSS attacks, on the importance of the ontology, and its core meaning for cybersecurity. We explain how a vulnerable website can be exploited, and how different JavaScript payloads can be used to detect vulnerabilities. We also enumerate some tools to use for an efficient analysis. We present detailed reasoning on what can be done to improve the security of a website in order to resist attacks, and we provide supportive examples. Then, we apply an ontology model against XSS attacks to strengthen the protection of a web application. However, we note that the existence of ontology does not improve the security itself, but it has to be properly used and should require a maximum of security layers to be taken into account.

Keywords: cybersecurity, web application vulnerabilities, cyber threats, ontology model

Procedia PDF Downloads 172
2140 Recent Advances in Data Warehouse

Authors: Fahad Hanash Alzahrani

Abstract:

This paper describes some recent advances in a quickly developing area of data storing and processing based on Data Warehouses and Data Mining techniques, which are associated with software, hardware, data mining algorithms and visualisation techniques having common features for any specific problems and tasks of their implementation.

Keywords: data warehouse, data mining, knowledge discovery in databases, on-line analytical processing

Procedia PDF Downloads 404