Search results for: Verification and Validation (V&V)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1886

Search results for: Verification and Validation (V&V)

836 On the Role of Cutting Conditions on Surface Roughness in High-Speed Thread Milling of Brass C3600

Authors: Amir Mahyar Khorasani, Ian Gibson, Moshe Goldberg, Mohammad Masoud Movahedi, Guy Littlefair

Abstract:

One of the important factors in manufacturing processes especially machining operations is surface quality. Improving this parameter results in improving fatigue strength, corrosion resistance, creep life and surface friction. The reliability and clearance of removable joints such as thread and nuts are highly related to the surface roughness. In this work, the effect of different cutting parameters such as cutting fluid pressure, feed rate and cutting speed on the surface quality of the crest of thread in the high-speed milling of Brass C3600 have been determined. Two popular neural networks containing MLP and RBF coupling with Taguchi L32 have been used to model surface roughness which was shown to be highly adept for such tasks. The contribution of this work is modelling surface roughness on the crest of the thread by using precise profilometer with nanoscale resolution. Experimental tests have been carried out for validation and approved suitable accuracy of the proposed model. Also analysing the interaction of parameters two by two showed that the most effective cutting parameter on the surface value is feed rate followed by cutting speed and cutting fluid pressure.

Keywords: artificial neural networks, cutting conditions, high-speed machining, surface roughness, thread milling

Procedia PDF Downloads 380
835 Shaping and Improving the Human Resource Management in Small and Medium Enterprises in Poland

Authors: Małgorzata Smolarek

Abstract:

One of the barriers to the development of small and medium-sized enterprises (SME) are difficulties connected with management of human resources. The first part of article defines the specifics of staff management in small and medium enterprises. The practical part presents results of own studies in the area of diagnosis of the state of the human resources management in small and medium-sized enterprises in Poland. It takes into account its impact on the functioning of SME in a variable environment. This part presents findings of empirical studies, which enabled verification of the hypotheses and formulation of conclusions. The findings presented in this paper were obtained during the implementation of the project entitled 'Tendencies and challenges in strategic managing SME in Silesian Voivodeship.' The aim of the studies was to diagnose the state of strategic management and human resources management taking into account its impact on the functioning of small and medium enterprises operating in Silesian Voivodeship in Poland and to indicate improvement areas of the model under diagnosis. One of the specific objectives of the studies was to diagnose the state of the process of strategic management of human resources and to identify fundamental problems. In this area, the main hypothesis was formulated: The enterprises analysed do not have comprehensive strategies for management of human resources. The survey was conducted by questionnaire. Main Research Results: Human resource management in SMEs is characterized by simplicity of procedures, and the lack of sophisticated tools and its specificity depends on the size of the company. The process of human resources management in SME has to be adjusted to the structure of an organisation, result from its objectives, so that an organisation can fully implement its strategic plans and achieve success and competitive advantage on the market. A guarantee of success is an accurately developed policy of human resources management based on earlier analyses of the existing procedures and possessed human resources.

Keywords: human resources management, human resources policy, personnel strategy, small and medium enterprises

Procedia PDF Downloads 243
834 Assessment the Correlation of Rice Yield Traits by Simulation and Modelling Methods

Authors: Davood Barari Tari

Abstract:

In order to investigate the correlation of rice traits in different nitrogen management methods by modeling programming, an experiment was laid out in rice paddy field in an experimental field at Caspian Coastal Sea region from 2013 to 2014. Variety used was Shiroudi as a high yielding variety. Nitrogen management was in two methods. Amount of nitrogen at four levels (30, 60, 90, and 120 Kg N ha-1 and control) and nitrogen-splitting at four levels (T1: 50% in base + 50% in maximum tillering stage, T2= 33.33% basal +33.33% in maximum tillering stage +33.33% in panicle initiation stage, T3=25% basal+37.5% in maximum tillering stage +37.5% in panicle initiation stage, T4: 25% in basal + 25% in maximum tillering stage + 50% in panicle initiation stage). Results showed that nitrogen traits, total grain number, filled spikelets, panicle number per m2 had a significant correlation with grain yield. Results related to calibrated and validation of rice model methods indicated that correlation between rice yield and yield components was accurate. The correlation between panicle length and grain yield was minimum. Physiological indices was simulated with low accuracy. According to results, investigation of the correlation between rice traits in physiological, morphological and phenological characters and yield by modeling and simulation methods are very useful.

Keywords: rice, physiology, modelling, simulation, yield traits

Procedia PDF Downloads 344
833 Arbitrarily Shaped Blur Kernel Estimation for Single Image Blind Deblurring

Authors: Aftab Khan, Ashfaq Khan

Abstract:

The research paper focuses on an interesting challenge faced in Blind Image Deblurring (BID). It relates to the estimation of arbitrarily shaped or non-parametric Point Spread Functions (PSFs) of motion blur caused by camera handshake. These PSFs exhibit much more complex shapes than their parametric counterparts and deblurring in this case requires intricate ways to estimate the blur and effectively remove it. This research work introduces a novel blind deblurring scheme visualized for deblurring images corrupted by arbitrarily shaped PSFs. It is based on Genetic Algorithm (GA) and utilises the Blind/Reference-less Image Spatial QUality Evaluator (BRISQUE) measure as the fitness function for arbitrarily shaped PSF estimation. The proposed BID scheme has been compared with other single image motion deblurring schemes as benchmark. Validation has been carried out on various blurred images. Results of both benchmark and real images are presented. Non-reference image quality measures were used to quantify the deblurring results. For benchmark images, the proposed BID scheme using BRISQUE converges in close vicinity of the original blurring functions.

Keywords: blind deconvolution, blind image deblurring, genetic algorithm, image restoration, image quality measures

Procedia PDF Downloads 444
832 Numerical Analysis of Shear Crack Propagation in a Concrete Beam without Transverse Reinforcement

Authors: G. A. Rombach, A. Faron

Abstract:

Crack formation and growth in reinforced concrete members are, in many cases, the cause of the collapse of technical structures. Such serious failures impair structural behavior and can also damage property and persons. An intensive investigation of the crack propagation is indispensable. Numerical methods are being developed to analyze crack growth in an element and to detect fracture failure at an early stage. For reinforced concrete components, however, further research and action are required in the analysis of shear cracks. This paper presents numerical simulations and continuum mechanical modeling of bending shear crack propagation in a three-dimensional reinforced concrete beam without transverse reinforcement. The analysis will provide a further understanding of crack growth and redistribution of inner forces in concrete members. As a numerical method to map discrete cracks, the extended finite element method (XFEM) is applied. The crack propagation is compared with the smeared crack approach using concrete damage plasticity. For validation, the crack patterns of real experiments are compared with the results of the different finite element models. The evaluation is based on single span beams under bending. With the analysis, it is possible to predict the fracture behavior of concrete members.

Keywords: concrete damage plasticity, crack propagation, extended finite element method, fracture mechanics

Procedia PDF Downloads 120
831 Simultaneous Analysis of 25 Trace Elements in Micro Volume of Human Serum by Inductively Coupled Plasma–Mass Spectrometry

Authors: Azmawati Mohammed Nawi, Siok-Fong Chin, Shamsul Azhar Shah, Rahman Jamal

Abstract:

In recent years, trace elements have gained importance as biomarkers in many chronic diseases. Unfortunately, the requirement for sample volume increases according to the extent of investigation for diagnosis or elucidating the mechanism of the disease. Here, we describe the method development and validation for simultaneous determination of 25 trace elements (lithium (Li), beryllium (Be), magnesium (Mg), aluminium (Al), vanadium (V), chromium (Cr), manganese (Mn), iron (Fe), cobalt (Co), nickel (Ni), copper (Cu), zinc (Zn), gallium (Ga), arsenic (As), selenium (Se), rubidium (Rb), strontium (Sr), silver (Ag), cadmium (Cd), caesium (Cs), barium (Ba), mercury (Hg), thallium (Tl), lead (Pb), uranium (U)) using just 20 µL of human serum. Serum samples were digested with nitric acid and hydrochloric acid (ratio 1:1, v/v) and analysed using inductively coupled plasma–mass spectrometry (ICP-MS). Seronorm®, a human-derived serum control material was used as quality control samples. The intra-day and inter-day precisions were consistently < 15% for all elements. The validated method was later applied to 30 human serum samples to evaluate its suitability. In conclusion, we have successfully developed and validated a precise and accurate analytical method for determining 25 trace elements requiring very low volume of human serum.

Keywords: acid digestion, ICP-MS, trace element, serum

Procedia PDF Downloads 185
830 Brain Connectome of Glia, Axons, and Neurons: Cognitive Model of Analogy

Authors: Ozgu Hafizoglu

Abstract:

An analogy is an essential tool of human cognition that enables connecting diffuse and diverse systems with physical, behavioral, principal relations that are essential to learning, discovery, and innovation. The Cognitive Model of Analogy (CMA) leads and creates patterns of pathways to transfer information within and between domains in science, just as happens in the brain. The connectome of the brain shows how the brain operates with mental leaps between domains and mental hops within domains and the way how analogical reasoning mechanism operates. This paper demonstrates the CMA as an evolutionary approach to science, technology, and life. The model puts forward the challenges of deep uncertainty about the future, emphasizing the need for flexibility of the system in order to enable reasoning methodology to adapt to changing conditions in the new era, especially post-pandemic. In this paper, we will reveal how to draw an analogy to scientific research to discover new systems that reveal the fractal schema of analogical reasoning within and between the systems like within and between the brain regions. Distinct phases of the problem-solving processes are divided thusly: stimulus, encoding, mapping, inference, and response. Based on the brain research so far, the system is revealed to be relevant to brain activation considering each of these phases with an emphasis on achieving a better visualization of the brain’s mechanism in macro context; brain and spinal cord, and micro context: glia and neurons, relative to matching conditions of analogical reasoning and relational information, encoding, mapping, inference and response processes, and verification of perceptual responses in four-term analogical reasoning. Finally, we will relate all these terminologies with these mental leaps, mental maps, mental hops, and mental loops to make the mental model of CMA clear.

Keywords: analogy, analogical reasoning, brain connectome, cognitive model, neurons and glia, mental leaps, mental hops, mental loops

Procedia PDF Downloads 165
829 Improving Security Features of Traditional Automated Teller Machines-Based Banking Services via Fingerprint Biometrics Scheme

Authors: Anthony I. Otuonye, Juliet N. Odii, Perpetual N. Ibe

Abstract:

The obvious challenges faced by most commercial bank customers while using the services of ATMs (Automated Teller Machines) across developing countries have triggered the need for an improved system with better security features. Current ATM systems are password-based, and research has proved the vulnerabilities of these systems to heinous attacks and manipulations. We have discovered by research that the security of current ATM-assisted banking services in most developing countries of the world is easily broken and maneuvered by fraudsters, majorly because it is quite difficult for these systems to identify an impostor with privileged access as against the authentic bank account owner. Again, PIN (Personal Identification Number) code passwords are easily guessed, just to mention a few of such obvious limitations of traditional ATM operations. In this research work also, we have developed a system of fingerprint biometrics with PIN code Authentication that seeks to improve the security features of traditional ATM installations as well as other Banking Services. The aim is to ensure better security at all ATM installations and raise the confidence of bank customers. It is hoped that our system will overcome most of the challenges of the current password-based ATM operation if properly applied. The researchers made use of the OOADM (Object-Oriented Analysis and Design Methodology), a software development methodology that assures proper system design using modern design diagrams. Implementation and coding were carried out using Visual Studio 2010 together with other software tools. Results obtained show a working system that provides two levels of security at the client’s side using a fingerprint biometric scheme combined with the existing 4-digit PIN code to guarantee the confidence of bank customers across developing countries.

Keywords: fingerprint biometrics, banking operations, verification, ATMs, PIN code

Procedia PDF Downloads 45
828 Development and Validation of an Electronic Module in Linear Motion for First Year College Students of Iloilo City

Authors: Donna H. Gabor

Abstract:

This study aimed to develop and validate an electronic module in physics for first-year college students of Iloilo and find out if there would be a significant difference in the performance of students before and after using the electronic module. The e-module was composed of one topic with two sub-lessons in linear motion (kinematics). The participants of the study were classified into three groups: the subject matter experts who are physics instructors who suggested the content, physical appearance, and limitations of the e-module; the IT experts who are active both in teaching and developing computer programs; and 28 students divided into two groups, 15 in the pilot group and 13 in the final test group. A researcher created 30 items checklist form (difficulty of a sample problem, comprehension, application, and definition of terms) was prepared and validated by the experts in subject matter for gathering data. To test the difference in student performance in physics, the researcher prepared an achievement test containing 25 items, multiple choices. The findings revealed that there was an increase in the performance of students in the pretest and post-test. T-test results revealed that there was a significant difference in the test scores of the students before and after using the module which can be used as a future reference for linear motion as an additional teaching tool in physics.

Keywords: electronic module, kinematics, linear motion, physics

Procedia PDF Downloads 135
827 Development and Validation of the Response to Stressful Situations Scale in the General Population

Authors: Célia Barreto Carvalho, Carolina da Motta, Marina Sousa, Joana Cabral, Ana Luísa Carvalho, Ermelindo Peixoto

Abstract:

The aim of the current study was to develop and validate a Response to Stressful Situations Scale (RSSS) for the Portuguese population. This scale assesses the degree of stress experienced in scenarios that can constitute positive, negative and more neutral stressors, and also describes the physiological, emotional and behavioral reactions to those events according to their intensity. These scenario include typical stressor scenarios relevant to patients with schizophrenia, which are currently absent from most scale, assessing specific risks that these stressors may bring on subjects, which may prove useful in non-clinical and clinical populations (i.e. patients with mood or anxiety disorders, schizophrenia). Results from Principal Components Analysis and Confirmatory Factor Analysis of on two adult samples from general population allowed to confirm a three-factor model with good fit indices: χ2 (144)= 370.211, p = 0.000; GFI = 0.928; CFI = 0.927; TLI = 0.914, RMSEA = 0.055, P( rmsea ≤ 0.005) = 0.096; PCFI = 0.781. Further data analysis on the scale revealed that RSSS is an adequate assessment tool of stress response in adults to be used in further research and clinical settings, with good psychometric characteristics, adequate divergent and convergent validity, good temporal stability and high internal consistency.

Keywords: assessment, stress events, stress response, stress vulnerability

Procedia PDF Downloads 520
826 Roasting Degree of Cocoa Beans by Artificial Neural Network (ANN) Based Electronic Nose System and Gas Chromatography (GC)

Authors: Juzhong Tan, William Kerr

Abstract:

Roasting is one critical procedure in chocolate processing, where special favors are developed, moisture content is decreased, and better processing properties are developed. Therefore, determination of roasting degree of cocoa bean is important for chocolate manufacturers to ensure the quality of chocolate products, and it also decides the commercial value of cocoa beans collected from cocoa farmers. The roasting degree of cocoa beans currently relies on human specialists, who sometimes are biased, and chemical analysis, which take long time and are inaccessible to many manufacturers and farmers. In this study, a self-made electronic nose system consists of gas sensors (TGS 800 and 2000 series) was used to detecting the gas generated by cocoa beans with a different roasting degree (0min, 20min, 30min, and 40min) and the signals collected by gas sensors were used to train a three-layers ANN. Chemical analysis of the graded beans was operated by traditional GC-MS system and the contents of volatile chemical compounds were used to train another ANN as a reference to electronic nosed signals trained ANN. Both trained ANN were used to predict cocoa beans with a different roasting degree for validation. The best accuracy of grading achieved by electronic nose signals trained ANN (using signals from TGS 813 826 820 880 830 2620 2602 2610) turned out to be 96.7%, however, the GC trained ANN got the accuracy of 83.8%.

Keywords: artificial neutron network, cocoa bean, electronic nose, roasting

Procedia PDF Downloads 235
825 Modeling Optimal Lipophilicity and Drug Performance in Ligand-Receptor Interactions: A Machine Learning Approach to Drug Discovery

Authors: Jay Ananth

Abstract:

The drug discovery process currently requires numerous years of clinical testing as well as money just for a single drug to earn FDA approval. For drugs that even make it this far in the process, there is a very slim chance of receiving FDA approval, resulting in detrimental hurdles to drug accessibility. To minimize these inefficiencies, numerous studies have implemented computational methods, although few computational investigations have focused on a crucial feature of drugs: lipophilicity. Lipophilicity is a physical attribute of a compound that measures its solubility in lipids and is a determinant of drug efficacy. This project leverages Artificial Intelligence to predict the impact of a drug’s lipophilicity on its performance by accounting for factors such as binding affinity and toxicity. The model predicted lipophilicity and binding affinity in the validation set with very high R² scores of 0.921 and 0.788, respectively, while also being applicable to a variety of target receptors. The results expressed a strong positive correlation between lipophilicity and both binding affinity and toxicity. The model helps in both drug development and discovery, providing every pharmaceutical company with recommended lipophilicity levels for drug candidates as well as a rapid assessment of early-stage drugs prior to any testing, eliminating significant amounts of time and resources currently restricting drug accessibility.

Keywords: drug discovery, lipophilicity, ligand-receptor interactions, machine learning, drug development

Procedia PDF Downloads 111
824 Validation of Modern Work Modules and Their Impact on Sustainable Human Resource Management in the Construction Industry

Authors: Robin Becker, Nane Roetmann, Manfred Helmus

Abstract:

The construction industry faces a significant challenge due to a shortage of skilled work-ers, especially in construction management, despite an increase in graduates. This is main-ly because the job is associated with high stress, long hours, and poor work-life balance. A survey revealed that the profession is unattractive to students, who prioritize personal growth, flexibility, and digitalization in their careers. To address this issue, companies can consider implementing various work modules like "working time documentation," "home office," "job sharing," and "time off." These modules can improve control, work-life bal-ance, and efficiency if tailored to the company's framework. They offer a way to make the field more appealing to future employees while benefiting existing staff, provided that both employers and employees are flexible and considerate of project-specific conditions and teams. The feasibility of these models depends on the company's overall framework, with potential for cost-neutral implementation and positive effects on efficiency and men-tal health. However, their success also relies on the specific context of the company, and more data is needed to assess their full impact.

Keywords: modern construction management, construction industry, work modules, shortage of junior staff, sustainable personnel management, making construction management more attractive, working time model

Procedia PDF Downloads 41
823 Comparison of Automated Zone Design Census Output Areas with Existing Output Areas in South Africa

Authors: T. Mokhele, O. Mutanga, F. Ahmed

Abstract:

South Africa is one of the few countries that have stopped using the same Enumeration Areas (EAs) for census enumeration and dissemination. The advantage of this change is that confidentiality issue could be addressed for census dissemination as the design of geographic unit for collection is mainly to ensure that this unit is covered by one enumerator. The objective of this paper was to evaluate the performance of automated zone design output areas against non-zone design developed geographies using the 2001 census data, and 2011 census to some extent, as the main input. The comparison of the Automated Zone-design Tool (AZTool) census output areas with the Small Area Layers (SALs) and SubPlaces based on confidentiality limit, population distribution, and degree of homogeneity, as well as shape compactness, was undertaken. Further, SPSS was employed for validation of the AZTool output results. The results showed that AZTool developed output areas out-perform the existing official SAL and SubPlaces with regard to minimum population threshold, population distribution and to some extent to homogeneity. Therefore, it was concluded that AZTool program provides a new alternative to the creation of optimised census output areas for dissemination of population census data in South Africa.

Keywords: AZTool, enumeration areas, small areal layers, South Africa

Procedia PDF Downloads 185
822 Comparison between Pushover Analysis Techniques and Validation of the Simplified Modal Pushover Analysis

Authors: N. F. Hanna, A. M. Haridy

Abstract:

One of the main drawbacks of the Modal Pushover Analysis (MPA) is the need to perform nonlinear time-history analysis, which complicates the analysis method and time. A simplified version of the MPA has been proposed based on the concept of the inelastic deformation ratio. Furthermore, the effect of the higher modes of vibration is considered by assuming linearly-elastic responses, which enables the use of standard elastic response spectrum analysis. In this thesis, the simplified MPA (SMPA) method is applied to determine the target global drift and the inter-story drifts of steel frame building. The effect of the higher vibration modes is considered within the framework of the SMPA. A comprehensive survey about the inelastic deformation ratio is presented. After that, a suitable expression from literature is selected for the inelastic deformation ratio and then implemented in the SMPA. The estimated seismic demands using the SMPA, such as target drift, base shear, and the inter-story drifts, are compared with the seismic responses determined by applying the standard MPA. The accuracy of the estimated seismic demands is validated by comparing with the results obtained by the nonlinear time-history analysis using real earthquake records.

Keywords: modal analysis, pushover analysis, seismic performance, target displacement

Procedia PDF Downloads 362
821 Evaluation of Golden Beam Data for the Commissioning of 6 and 18 MV Photons Beams in Varian Linear Accelerator

Authors: Shoukat Ali, Abdul Qadir Jandga, Amjad Hussain

Abstract:

Objective: The main purpose of this study is to compare the Percent Depth dose (PDD) and In-plane and cross-plane profiles of Varian Golden beam data to the measured data of 6 and 18 MV photons for the commissioning of Eclipse treatment planning system. Introduction: Commissioning of treatment planning system requires an extensive acquisition of beam data for the clinical use of linear accelerators. Accurate dose delivery require to enter the PDDs, Profiles and dose rate tables for open and wedges fields into treatment planning system, enabling to calculate the MUs and dose distribution. Varian offers a generic set of beam data as a reference data, however not recommend for clinical use. In this study, we compared the generic beam data with the measured beam data to evaluate the reliability of generic beam data to be used for the clinical purpose. Methods and Material: PDDs and Profiles of Open and Wedge fields for different field sizes and at different depths measured as per Varian’s algorithm commissioning guideline. The measurement performed with PTW 3D-scanning water phantom with semi-flex ion chamber and MEPHYSTO software. The online available Varian Golden Beam Data compared with the measured data to evaluate the accuracy of the golden beam data to be used for the commissioning of Eclipse treatment planning system. Results: The deviation between measured vs. golden beam data was in the range of 2% max. In PDDs, the deviation increases more in the deeper depths than the shallower depths. Similarly, profiles have the same trend of increasing deviation at large field sizes and increasing depths. Conclusion: Study shows that the percentage deviation between measured and golden beam data is within the acceptable tolerance and therefore can be used for the commissioning process; however, verification of small subset of acquired data with the golden beam data should be mandatory before clinical use.

Keywords: percent depth dose, flatness, symmetry, golden beam data

Procedia PDF Downloads 490
820 On the Use of Machine Learning for Tamper Detection

Authors: Basel Halak, Christian Hall, Syed Abdul Father, Nelson Chow Wai Kit, Ruwaydah Widaad Raymode

Abstract:

The attack surface on computing devices is becoming very sophisticated, driven by the sheer increase of interconnected devices, reaching 50B in 2025, which makes it easier for adversaries to have direct access and perform well-known physical attacks. The impact of increased security vulnerability of electronic systems is exacerbated for devices that are part of the critical infrastructure or those used in military applications, where the likelihood of being targeted is very high. This continuously evolving landscape of security threats calls for a new generation of defense methods that are equally effective and adaptive. This paper proposes an intelligent defense mechanism to protect from physical tampering, it consists of a tamper detection system enhanced with machine learning capabilities, which allows it to recognize normal operating conditions, classify known physical attacks and identify new types of malicious behaviors. A prototype of the proposed system has been implemented, and its functionality has been successfully verified for two types of normal operating conditions and further four forms of physical attacks. In addition, a systematic threat modeling analysis and security validation was carried out, which indicated the proposed solution provides better protection against including information leakage, loss of data, and disruption of operation.

Keywords: anti-tamper, hardware, machine learning, physical security, embedded devices, ioT

Procedia PDF Downloads 154
819 Jungle Justice on Emotional Health Challenges among Lagosians

Authors: Aaron Akinloye

Abstract:

This research examined the influence of jungle justice as it affects the emotional health challenges among residents in Lagos metropolitan city. Descriptive survey research design was used along with the questionnaire as research instrument. Population for the study comprised residents in Yaba and Shomolu Communities of Lagos State, Nigeria. Accidental sampling technique was used to sample 300 Residents. Self-developed questionnaire was used to obtain data on the variables under investigation. Research instrument was validated following the face, content, and construct validation of the instrument. Thereafter, the reliability coefficient yielded 0.84. It is therefore concluded and recommended that; there is a significant influence of jungle justice on trauma among residents- df (298) t= 2.33, p< 0.05; there is a significant influence of jungle justice on pressure among residents- df (298) t= 2.16, p< 0.05: there is a significant influence of jungle justice on fear among residents- df (298) t= 2.20, p< 0.05; there is a significant influence of jungle justice on depression among residents- df (298) t= 2.14, p< 0.05. Recommendations were made that; there should be deliberate effort to implement comprehensive awareness campaigns to educate the residents on the detrimental effects of jungle justice on individuals and the community members as a whole; there should be an improvement in the effectiveness and efficiency of the existing law enforcement agencies in Lagos metropolitan city; development and implementation of support systems for victims of jungle justice, which include trauma, counselling, mental health services, and rehabilitation programmes; there should be proper review and revision of the legal framework to address the issue of jungle justice effectively.

Keywords: jungle justice, emotional health, depression, fear

Procedia PDF Downloads 101
818 The Role of the Elastic Foundation Having Nonlinear Stiffness Properties in the Vibration of Structures

Authors: E. Feulefack Songong, A. Zingoni

Abstract:

A vibration is a mechanical phenomenon whereby oscillations occur about an equilibrium point. Although vibrations can be linear or nonlinear depending on the basic components of the system, the interest is mostly pointed towards nonlinear vibrations. This is because most structures around us are to some extent nonlinear and also because we need more accurate values in an analysis. The goal of this research is the integration of nonlinearities in the development and validation of structural models and to ameliorate the resistance of structures when subjected to loads. Although there exist many types of nonlinearities, this thesis will mostly focus on the vibration of free and undamped systems incorporating nonlinearity due to stiffness. Nonlinear stiffness has been a concern to many engineers in general and Civil engineers in particular because it is an important factor that can bring a good modification and amelioration to the response of structures when subjected to loads. The analysis of systems will be done analytically and then numerically to validate the analytical results. We will first show the benefit and importance of stiffness nonlinearity when it is implemented in the structure. Secondly, We will show how its integration in the structure can improve not only the structure’s performance but also its response when subjected to loads. The results of this study will be valuable to practicing engineers as well as industry practitioners in developing better designs and tools for their structures and mechanical devices. They will also serve to engineers to design lighter and stronger structures and to give good predictions as for the behavior of structures when subjected to external loads.

Keywords: elastic foundation, nonlinear, plates, stiffness, structures, vibration

Procedia PDF Downloads 136
817 Catchment Yield Prediction in an Ungauged Basin Using PyTOPKAPI

Authors: B. S. Fatoyinbo, D. Stretch, O. T. Amoo, D. Allopi

Abstract:

This study extends the use of the Drainage Area Regionalization (DAR) method in generating synthetic data and calibrating PyTOPKAPI stream yield for an ungauged basin at a daily time scale. The generation of runoff in determining a river yield has been subjected to various topographic and spatial meteorological variables, which integers form the Catchment Characteristics Model (CCM). Many of the conventional CCM models adapted in Africa have been challenged with a paucity of adequate, relevance and accurate data to parameterize and validate the potential. The purpose of generating synthetic flow is to test a hydrological model, which will not suffer from the impact of very low flows or very high flows, thus allowing to check whether the model is structurally sound enough or not. The employed physically-based, watershed-scale hydrologic model (PyTOPKAPI) was parameterized with GIS-pre-processing parameters and remote sensing hydro-meteorological variables. The validation with mean annual runoff ratio proposes a decent graphical understanding between observed and the simulated discharge. The Nash-Sutcliffe efficiency and coefficient of determination (R²) values of 0.704 and 0.739 proves strong model efficiency. Given the current climate variability impact, water planner can now assert a tool for flow quantification and sustainable planning purposes.

Keywords: catchment characteristics model, GIS, synthetic data, ungauged basin

Procedia PDF Downloads 327
816 Verification of a Simple Model for Rolling Isolation System Response

Authors: Aarthi Sridhar, Henri Gavin, Karah Kelly

Abstract:

Rolling Isolation Systems (RISs) are simple and effective means to mitigate earthquake hazards to equipment in critical and precious facilities, such as hospitals, network collocation facilities, supercomputer centers, and museums. The RIS works by isolating components acceleration the inertial forces felt by the subsystem. The RIS consists of two platforms with counter-facing concave surfaces (dishes) in each corner. Steel balls lie inside the dishes and allow the relative motion between the top and bottom platform. Formerly, a mathematical model for the dynamics of RISs was developed using Lagrange’s equations (LE) and experimentally validated. A new mathematical model was developed using Gauss’s Principle of Least Constraint (GPLC) and verified by comparing impulse response trajectories of the GPLC model and the LE model in terms of the peak displacements and accelerations of the top platform. Mathematical models for the RIS are tedious to derive because of the non-holonomic rolling constraints imposed on the system. However, using Gauss’s Principle of Least constraint to find the equations of motion removes some of the obscurity and yields a system that can be easily extended. Though the GPLC model requires more state variables, the equations of motion are far simpler. The non-holonomic constraint is enforced in terms of accelerations and therefore requires additional constraint stabilization methods in order to avoid the possibility that numerical integration methods can cause the system to go unstable. The GPLC model allows the incorporation of more physical aspects related to the RIS, such as contribution of the vertical velocity of the platform to the kinetic energy and the mass of the balls. This mathematical model for the RIS is a tool to predict the motion of the isolation platform. The ability to statistically quantify the expected responses of the RIS is critical in the implementation of earthquake hazard mitigation.

Keywords: earthquake hazard mitigation, earthquake isolation, Gauss’s Principle of Least Constraint, nonlinear dynamics, rolling isolation system

Procedia PDF Downloads 252
815 Lifelong Education for Teachers: A Tool for Achieving Effective Teaching and Learning in Secondary Schools in Benue State, Nigeria

Authors: Adzongo Philomena Ibuh, Aloga O. Austin

Abstract:

The purpose of the study was to examine lifelong education for teachers as a tool for achieving effective teaching and learning. Lifelong education enhances social inclusion, personal development, citizenship, employability, teaching and learning, community and the nation, and the challenges of lifelong education were also discussed. Descriptive survey design was adopted for the study. A simple random sampling technique was used to select 80 teachers as sample from a population of 105 senior secondary school teachers in Makurdi local government area of Benue state. A 20-item self designed questionnaire subjected to expert validation and reliability was used to collect data. The reliability Alpha coefficient of 0.87 was established using Crombach Alpha technique, mean scores and standard deviation were used to answer the 2 research questions while chi-square was used to analyze data for the 2 hypotheses. The findings of the study revealed that, lifelong education for teachers can be used to achieve as a tool for achieving effective teaching and learning, and the study recommended among others that government, organizations and individuals should in collaboration put lifelong education programmes for teachers on the priority list. The paper concluded that the strategic position of lifelong education for teachers towards enhanced teaching and learning makes it imperative for all hands to be on deck to support the programme financially and otherwise.

Keywords: effective teaching and learning, lifelong education, teachers, tool

Procedia PDF Downloads 476
814 ELISA Based hTSH Assessment Using Two Sensitive and Specific Anti-hTSH Polyclonal Antibodies

Authors: Maysam Mard-Soltani, Mohamad Javad Rasaee, Saeed Khalili, Abdol Karim Sheikhi, Mehdi Hedayati

Abstract:

Production of specific antibody responses against hTSH is a cumbersome process due to the high identity between the hTSH and the other members of the glycoprotein hormone family (FSH, LH and HCG) and the high identity between the human hTSH and host animals for antibody production. Therefore, two polyclonal antibodies were purified against two recombinant proteins. Four possible ELISA tests were designed based on these antibodies. These ELISA tests were checked against hTSH and other glycoprotein hormones, and their sensitivity and specificity were assessed. Bioinformatics tools were used to analyze the immunological properties. After the immunogen region selection from hTSH protein, c terminal of B hTSH was selected and applied. Two recombinant genes, with these cut pieces (first: two repeats of C terminal of B hTSH, second: tetanous toxin+B hTSH C terminal), were designed and sub-cloned into the pET32a expression vector. Standard methods were used for protein expression, purification, and verification. Thereafter, immunizations of the white New Zealand rabbits were performed and the serums of them were used for antibody titration, purification and characterization. Then, four ELISA tests based on two antibodies were employed to assess the hTSH and other glycoprotein hormones. The results of these assessments were compared with standard amounts. The obtained results indicated that the desired antigens were successfully designed, sub-cloned, expressed, confirmed and used for in vivo immunization. The raised antibodies were capable of specific and sensitive hTSH detection, while the cross reactivity with the other members of the glycoprotein hormone family was minimum. Among the four designed tests, the test in which the antibody against first protein was used as capture antibody, and the antibody against second protein was used as detector antibody did not show any hook effect up to 50 miu/l. Both proteins have the ability to induce highly sensitive and specific antibody responses against the hTSH. One of the antibody combinations of these antibodies has the highest sensitivity and specificity in hTSH detection.

Keywords: hTSH, bioinformatics, protein expression, cross reactivity

Procedia PDF Downloads 190
813 Insecurity as a Challenge to Nutritional Status of Children and Mothers in Dansadau, Maru Local Government Area Zamfara State, North Western Nigeria

Authors: Mohammed Hussaini

Abstract:

This paper discusses insecurity as a challenge to the nutritional Status of children and mothers in Dansadau, Maru Local Government area of Zamfara state, Northwestern Nigeria. A Descriptive survey design was used in the study. Objectives of the study were formulated to guide the study. 20 Health workers and 100 mothers were used as population of the study; the instrument validation for data collection was interview. The interview structure was validated by 3 experts, the data collected was analyzed and presented using descriptive standard score (Z-score). The study revealed that, Nutritional Status of children and mothers in Northwest Nigeria specifically Zamfara state is low. This mostly affect children and mother as a result of serious insecurity challenge in the region, consisting of banditry and kidnapping, killing of farmers, destruction of farmland, burning of farm products. The study recommended that the focus is on implementing strong communication strategies to enhance short-term relief initiatives, both governmental and non-governmental organizations should actively play a role in initiating lasting change, especially when tackling issues of insecurity and effectively addressing the rise of armed banditry and other security concerns requires a sophisticated and nuanced strategy.

Keywords: insecurity, malnutrition, children, mothers

Procedia PDF Downloads 55
812 Verification and Proposal of Information Processing Model Using EEG-Based Brain Activity Monitoring

Authors: Toshitaka Higashino, Naoki Wakamiya

Abstract:

Human beings perform a task by perceiving information from outside, recognizing them, and responding them. There have been various attempts to analyze and understand internal processes behind the reaction to a given stimulus by conducting psychological experiments and analysis from multiple perspectives. Among these, we focused on Model Human Processor (MHP). However, it was built based on psychological experiments and thus the relation with brain activity was unclear so far. To verify the validity of the MHP and propose our model from a viewpoint of neuroscience, EEG (Electroencephalography) measurements are performed during experiments in this study. More specifically, first, experiments were conducted where Latin alphabet characters were used as visual stimuli. In addition to response time, ERPs (event-related potentials) such as N100 and P300 were measured by using EEG. By comparing cycle time predicted by the MHP and latency of ERPs, it was found that N100, related to perception of stimuli, appeared at the end of the perceptual processor. Furthermore, by conducting an additional experiment, it was revealed that P300, related to decision making, appeared during the response decision process, not at the end. Second, by experiments using Japanese Hiragana characters, i.e. Japan's own phonetic symbols, those findings were confirmed. Finally, Japanese Kanji characters were used as more complicated visual stimuli. A Kanji character usually has several readings and several meanings. Despite the difference, a reading-related task and a meaning-related task exhibited similar results, meaning that they involved similar information processing processes of the brain. Based on those results, our model was proposed which reflects response time and ERP latency. It consists of three processors: the perception processor from an input of a stimulus to appearance of N100, the cognitive processor from N100 to P300, and the decision-action processor from P300 to response. Using our model, an application system which reflects brain activity can be established.

Keywords: brain activity, EEG, information processing model, model human processor

Procedia PDF Downloads 99
811 A Comparative Asessment of Some Algorithms for Modeling and Forecasting Horizontal Displacement of Ialy Dam, Vietnam

Authors: Kien-Trinh Thi Bui, Cuong Manh Nguyen

Abstract:

In order to simulate and reproduce the operational characteristics of a dam visually, it is necessary to capture the displacement at different measurement points and analyze the observed movement data promptly to forecast the dam safety. The accuracy of forecasts is further improved by applying machine learning methods to data analysis progress. In this study, the horizontal displacement monitoring data of the Ialy hydroelectric dam was applied to machine learning algorithms: Gaussian processes, multi-layer perceptron neural networks, and the M5-rules algorithm for modelling and forecasting of horizontal displacement of the Ialy hydropower dam (Vietnam), respectively, for analysing. The database which used in this research was built by collecting time series of data from 2006 to 2021 and divided into two parts: training dataset and validating dataset. The final results show all three algorithms have high performance for both training and model validation, but the MLPs is the best model. The usability of them are further investigated by comparison with a benchmark models created by multi-linear regression. The result show the performance which obtained from all the GP model, the MLPs model and the M5-Rules model are much better, therefore these three models should be used to analyze and predict the horizontal displacement of the dam.

Keywords: Gaussian processes, horizontal displacement, hydropower dam, Ialy dam, M5-Rules, multi-layer perception neural networks

Procedia PDF Downloads 213
810 Eliminating Arm, Neck and Leg Fatigue of United Asia International Plastics Corporation Workers through Rapid Entire Body Assessment

Authors: John Cheferson R. De Belen, John Paul G. Elizares, Ronald John G. Raz, Janina Elyse A. Reyes, Charie G. Salengua, Aristotle L. Soriano

Abstract:

Plastic is a type of synthetic or man-made polymer that can readily be molded into a variety of products. Its usage over the past century has enabled society to make huge technological advances. The workers of United Asia International Plastics Corporation (UAIPC), a plastic manufacturing company performs manual packaging which causes fatigue and stress on their arm, neck, and legs due to extended periods of standing and repetitive motions. With the use of the Fishbone Diagram, Five-Why Analysis, Rapid Entire Body Assessment (REBA), and Anthropometry, the stressful tasks and activities were identified and analyzed. Given the anthropometric measurements obtained from the workers, improved dimensions for the tables and chairs should be used and provide a new packaging machine. The validation of this proposal shall follow after its implementation. By eliminating fatigue during working hours in the production, the workers will be at ease at performing their work properly; productivity will increase that will lead to more profit. Further areas for study include measurement and comparison of the worker’s anthropometric measurement with the industry standard.

Keywords: anthropometry, fishbone diagram, five-why analysis, rapid entire body assessment

Procedia PDF Downloads 266
809 Geological Mapping of Gabel Humr Akarim Area, Southern Eastern Desert, Egypt: Constrain from Remote Sensing Data, Petrographic Description and Field Investigation

Authors: Doaa Hamdi, Ahmed Hashem

Abstract:

The present study aims at integrating the ASTER data and Landsat 8 data to discriminate and map alteration and/or mineralization zones in addition to delineating different lithological units of Humr Akarim Granites area. The study area is located at 24º9' to 24º13' N and 34º1' to 34º2'45"E., covering a total exposed surface area of about 17 km². The area is characterized by rugged topography with low to moderate relief. Geologic fieldwork and petrographic investigations revealed that the basement complex of the study area is composed of metasediments, mafic dikes, older granitoids, and alkali-feldspar granites. Petrographic investigations revealed that the secondary minerals in the study area are mainly represented by chlorite, epidote, clay minerals and iron oxides. These minerals have specific spectral signatures in the region of visible near-infrared and short-wave infrared (0.4 to 2.5 µm). So that the ASTER imagery processing was concentrated on VNIR-SWIR spectrometric data in order to achieve the purposes of this study (geologic mapping of hydrothermal alteration zones and delineate possible radioactive potentialities). Mapping of hydrothermal alterations zones in addition to discriminating the lithological units in the study area are achieved through the utilization of some different image processing, including color band composites (CBC) and data transformation techniques such as band ratios (BR), band ratio codes (BRCs), principal component analysis(PCA), Crosta Technique and minimum noise fraction (MNF). The field verification and petrographic investigation confirm the results of ASTER imagery and Landsat 8 data, proposing a geological map (scale 1:50000).

Keywords: remote sensing, petrography, mineralization, alteration detection

Procedia PDF Downloads 166
808 The Effect of Sorafenibe on Soat1 Protein by Using Molecular Docking Method

Authors: Mahdiyeh Gholaminezhad

Abstract:

Context: The study focuses on the potential impact of Sorafenib on SOAT1 protein in liver cancer treatment, addressing the need for more effective therapeutic options. Research aim: To explore the effects of Sorafenib on the activity of SOAT1 protein in liver cancer cells. Methodology: Molecular docking was employed to analyze the interaction between Sorafenib and SOAT1 protein. Findings: The study revealed a significant effect of Sorafenib on the stability and activity of SOAT1 protein, suggesting its potential as a treatment for liver cancer. Theoretical importance: This research highlights the molecular mechanism underlying Sorafenib's anti-cancer properties, contributing to the understanding of its therapeutic effects. Data collection: Data on the molecular structure of Sorafenib and SOAT1 protein were obtained from computational simulations and databases. Analysis procedures: Molecular docking simulations were performed to predict the binding interactions between Sorafenib and SOAT1 protein. Question addressed: How does Sorafenib influence the activity of SOAT1 protein and what are the implications for liver cancer treatment? Conclusion: The study demonstrates the potential of Sorafenib as a targeted therapy for liver cancer by affecting the activity of SOAT1 protein. Reviewers' Comments: The study provides valuable insights into the molecular basis of Sorafenib's action on SOAT1 protein, suggesting its therapeutic potential. To enhance the methodology, the authors could consider validating the docking results with experimental data for further validation.

Keywords: liver cancer, sorafenib, SOAT1, molecular docking

Procedia PDF Downloads 28
807 A Systemic Maturity Model

Authors: Emir H. Pernet, Jeimy J. Cano

Abstract:

Maturity models, used descriptively to explain changes in reality or normatively to guide managers to make interventions to make organizations more effective and efficient, are based on the principles of statistical quality control promulgated by Shewhart in the years 30, and on the principles of PDCA continuous improvement (Plan, Do, Check, Act) developed by Deming and Juran. Some frameworks developed over the concept of maturity models includes COBIT, CMM, and ITIL. This paper presents some limitations of traditional maturity models, most of them based on points of reflection and analysis done by some authors. Almost all limitations are related to the mechanistic and reductionist approach of the principles over those models are built. As Systems Theory helps the understanding of the dynamics of organizations and organizational change, the development of a systemic maturity model can help to overcome some of those limitations. This document proposes a systemic maturity model, based on a systemic conceptualization of organizations, focused on the study of the functioning of the parties, the relationships among them, and their behavior as a whole. The concept of maturity from the system theory perspective is conceptually defined as an emergent property of the organization, which arises from as a result of the degree of alignment and integration of their processes. This concept is operationalized through a systemic function that measures the maturity of an organization, and finally validated by the measuring of maturity in organizations. For its operationalization and validation, the model was applied to measure the maturity of organizational Governance, Risk and Compliance (GRC) processes.

Keywords: GRC, maturity model, systems theory, viable system model

Procedia PDF Downloads 312