Search results for: Artificial Bee Colony Algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5424

Search results for: Artificial Bee Colony Algorithm

624 A Context Aware Mobile Learning System with a Cognitive Recommendation Engine

Authors: Jalal Maqbool, Gyu Myoung Lee

Abstract:

Using smart devices for context aware mobile learning is becoming increasingly popular. This has led to mobile learning technology becoming an indispensable part of today’s learning environment and platforms. However, some fundamental issues remain - namely, mobile learning still lacks the ability to truly understand human reaction and user behaviour. This is due to the fact that current mobile learning systems are passive and not aware of learners’ changing contextual situations. They rely on static information about mobile learners. In addition, current mobile learning platforms lack the capability to incorporate dynamic contextual situations into learners’ preferences. Thus, this thesis aims to address these issues highlighted by designing a context aware framework which is able to sense learner’s contextual situations, handle data dynamically, and which can use contextual information to suggest bespoke learning content according to a learner’s preferences. This is to be underpinned by a robust recommendation system, which has the capability to perform these functions, thus providing learners with a truly context-aware mobile learning experience, delivering learning contents using smart devices and adapting to learning preferences as and when it is required. In addition, part of designing an algorithm for the recommendation engine has to be based on learner and application needs, personal characteristics and circumstances, as well as being able to comprehend human cognitive processes which would enable the technology to interact effectively and deliver mobile learning content which is relevant, according to the learner’s contextual situations. The concept of this proposed project is to provide a new method of smart learning, based on a capable recommendation engine for providing an intuitive mobile learning model based on learner actions.

Keywords: aware, context, learning, mobile

Procedia PDF Downloads 237
623 Near Optimal Closed-Loop Guidance Gains Determination for Vector Guidance Law, from Impact Angle Errors and Miss Distance Considerations

Authors: Karthikeyan Kalirajan, Ashok Joshi

Abstract:

An optimization problem is to setup to maximize the terminal kinetic energy of a maneuverable reentry vehicle (MaRV). The target location, the impact angle is given as constraints. The MaRV uses an explicit guidance law called Vector guidance. This law has two gains which are taken as decision variables. The problem is to find the optimal value of these gains which will result in minimum miss distance and impact angle error. Using a simple 3DOF non-rotating flat earth model and Lockheed martin HP-MARV as the reentry vehicle, the nature of solutions of the optimization problem is studied. This is achieved by carrying out a parametric study for a range of closed loop gain values and the corresponding impact angle error and the miss distance values are generated. The results show that there are well defined lower and upper bounds on the gains that result in near optimal terminal guidance solution. It is found from this study, that there exist common permissible regions (values of gains) where all constraints are met. Moreover, the permissible region lies between flat regions and hence the optimization algorithm has to be chosen carefully. It is also found that, only one of the gain values is independent and that the other dependent gain value is related through a simple straight-line expression. Moreover, to reduce the computational burden of finding the optimal value of two gains, a guidance law called Diveline guidance is discussed, which uses single gain. The derivation of the Diveline guidance law from Vector guidance law is discussed in this paper.

Keywords: Marv guidance, reentry trajectory, trajectory optimization, guidance gain selection

Procedia PDF Downloads 412
622 Development of a Geomechanical Risk Assessment Model for Underground Openings

Authors: Ali Mortazavi

Abstract:

The main objective of this research project is to delve into a multitude of geomechanical risks associated with various mining methods employed within the underground mining industry. Controlling geotechnical design parameters and operational factors affecting the selection of suitable mining techniques for a given underground mining condition will be considered from a risk assessment point of view. Important geomechanical challenges will be investigated as appropriate and relevant to the commonly used underground mining methods. Given the complicated nature of rock mass in-situ and complicated boundary conditions and operational complexities associated with various underground mining methods, the selection of a safe and economic mining operation is of paramount significance. Rock failure at varying scales within the underground mining openings is always a threat to mining operations and causes human and capital losses worldwide. Geotechnical design is a major design component of all underground mines and basically dominates the safety of an underground mine. With regard to uncertainties that exist in rock characterization prior to mine development, there are always risks associated with inappropriate design as a function of mining conditions and the selected mining method. Uncertainty often results from the inherent variability of rock masse, which in turn is a function of both geological materials and rock mass in-situ conditions. The focus of this research is on developing a methodology which enables a geomechanical risk assessment of given underground mining conditions. The outcome of this research is a geotechnical risk analysis algorithm, which can be used as an aid in selecting the appropriate mining method as a function of mine design parameters (e.g., rock in-situ properties, design method, governing boundary conditions such as in-situ stress and groundwater, etc.).

Keywords: geomechanical risk assessment, rock mechanics, underground mining, rock engineering

Procedia PDF Downloads 135
621 Resonant Fluorescence in a Two-Level Atom and the Terahertz Gap

Authors: Nikolai N. Bogolubov, Andrey V. Soldatov

Abstract:

Terahertz radiation occupies a range of frequencies somewhere from 100 GHz to approximately 10 THz, just between microwaves and infrared waves. This range of frequencies holds promise for many useful applications in experimental applied physics and technology. At the same time, reliable, simple techniques for generation, amplification, and modulation of electromagnetic radiation in this range are far from been developed enough to meet the requirements of its practical usage, especially in comparison to the level of technological abilities already achieved for other domains of the electromagnetic spectrum. This situation of relative underdevelopment of this potentially very important range of electromagnetic spectrum is known under the name of the 'terahertz gap.' Among other things, technological progress in the terahertz area has been impeded by the lack of compact, low energy consumption, easily controlled and continuously radiating terahertz radiation sources. Therefore, development of new techniques serving this purpose as well as various devices based on them is of obvious necessity. No doubt, it would be highly advantageous to employ the simplest of suitable physical systems as major critical components in these techniques and devices. The purpose of the present research was to show by means of conventional methods of non-equilibrium statistical mechanics and the theory of open quantum systems, that a thoroughly studied two-level quantum system, also known as an one-electron two-level 'atom', being driven by external classical monochromatic high-frequency (e.g. laser) field, can radiate continuously at much lower (e.g. terahertz) frequency in the fluorescent regime if the transition dipole moment operator of this 'atom' possesses permanent non-equal diagonal matrix elements. This assumption contradicts conventional assumption routinely made in quantum optics that only the non-diagonal matrix elements persist. The conventional assumption is pertinent to natural atoms and molecules and stems from the property of spatial inversion symmetry of their eigenstates. At the same time, such an assumption is justified no more in regard to artificially manufactured quantum systems of reduced dimensionality, such as, for example, quantum dots, which are often nicknamed 'artificial atoms' due to striking similarity of their optical properties to those ones of the real atoms. Possible ways to experimental observation and practical implementation of the predicted effect are discussed too.

Keywords: terahertz gap, two-level atom, resonant fluorescence, quantum dot, resonant fluorescence, two-level atom

Procedia PDF Downloads 258
620 Vibration Analysis of Stepped Nanoarches with Defects

Authors: Jaan Lellep, Shahid Mubasshar

Abstract:

A numerical solution is developed for simply supported nanoarches based on the non-local theory of elasticity. The nanoarch under consideration has a step-wise variable cross-section and is weakened by crack-like defects. It is assumed that the cracks are stationary and the mechanical behaviour of the nanoarch can be modeled by Eringen’s non-local theory of elasticity. The physical and thermal properties are sensitive with respect to changes of dimensions in the nano level. The classical theory of elasticity is unable to describe such changes in material properties. This is because, during the development of the classical theory of elasticity, the speculation of molecular objects was avoided. Therefore, the non-local theory of elasticity is applied to study the vibration of nanostructures and it has been accepted by many researchers. In the non-local theory of elasticity, it is assumed that the stress state of the body at a given point depends on the stress state of each point of the structure. However, within the classical theory of elasticity, the stress state of the body depends only on the given point. The system of main equations consists of equilibrium equations, geometrical relations and constitutive equations with boundary and intermediate conditions. The system of equations is solved by using the method of separation of variables. Consequently, the governing differential equations are converted into a system of algebraic equations whose solution exists if the determinant of the coefficients of the matrix vanishes. The influence of cracks and steps on the natural vibration of the nanoarches is prescribed with the aid of additional local compliance at the weakened cross-section. An algorithm to determine the eigenfrequencies of the nanoarches is developed with the help of computer software. The effects of various physical and geometrical parameters are recorded and drawn graphically.

Keywords: crack, nanoarches, natural frequency, step

Procedia PDF Downloads 121
619 Fatigue Life Prediction under Variable Loading Based a Non-Linear Energy Model

Authors: Aid Abdelkrim

Abstract:

A method of fatigue damage accumulation based upon application of energy parameters of the fatigue process is proposed in the paper. Using this model is simple, it has no parameter to be determined, it requires only the knowledge of the curve W–N (W: strain energy density N: number of cycles at failure) determined from the experimental Wöhler curve. To examine the performance of nonlinear models proposed in the estimation of fatigue damage and fatigue life of components under random loading, a batch of specimens made of 6082 T 6 aluminium alloy has been studied and some of the results are reported in the present paper. The paper describes an algorithm and suggests a fatigue cumulative damage model, especially when random loading is considered. This work contains the results of uni-axial random load fatigue tests with different mean and amplitude values performed on 6082T6 aluminium alloy specimens. The proposed model has been formulated to take into account the damage evolution at different load levels and it allows the effect of the loading sequence to be included by means of a recurrence formula derived for multilevel loading, considering complex load sequences. It is concluded that a ‘damaged stress interaction damage rule’ proposed here allows a better fatigue damage prediction than the widely used Palmgren–Miner rule, and a formula derived in random fatigue could be used to predict the fatigue damage and fatigue lifetime very easily. The results obtained by the model are compared with the experimental results and those calculated by the most fatigue damage model used in fatigue (Miner’s model). The comparison shows that the proposed model, presents a good estimation of the experimental results. Moreover, the error is minimized in comparison to the Miner’s model.

Keywords: damage accumulation, energy model, damage indicator, variable loading, random loading

Procedia PDF Downloads 384
618 Entropy in a Field of Emergence in an Aspect of Linguo-Culture

Authors: Nurvadi Albekov

Abstract:

Communicative situation is a basis, which designates potential models of ‘constructed forms’, a motivated basis of a text, for a text can be assumed as a product of the communicative situation. It is within the field of emergence the models of text, that can be potentially prognosticated in a certain communicative situation, are designated. Every text can be assumed as conceptual system structured on the base of certain communicative situation. However in the process of ‘structuring’ of a certain model of ‘conceptual system’ consciousness of a recipient is able act only within the border of the field of emergence for going out of this border indicates misunderstanding of the communicative situation. On the base of communicative situation we can witness the increment of meaning where the synergizing of the informative model of communication, formed by using of the invariant units of a language system, is a result of verbalization of the communicative situation. The potential of the models of a text, prognosticated within the field of emergence, also depends on the communicative situation. The conception ‘the field of emergence’ is interpreted as a unit of the language system, having poly-directed universal structure, implying the presence of the core, the center and the periphery, including different levels of means of a functioning system of language, both in terms of linguistic resources, and in terms of extra linguistic factors interaction of which results increment of a text. The conception ‘field of emergence’ is considered as the most promising in the analysis of texts: oral, written, printed and electronic. As a unit of the language system field of emergence has several properties that predict its use during the study of a text in different levels. This work is an attempt analysis of entropy in a text in the aspect of lingua-cultural code, prognosticated within the model of the field of emergence. The article describes the problem of entropy in the field of emergence, caused by influence of the extra-linguistic factors. The increasing of entropy is caused not only by the fact of intrusion of the language resources but by influence of the alien culture in a whole, and by appearance of non-typical for this very culture symbols in the field of emergence. The borrowing of alien lingua-cultural symbols into the lingua-culture of the author is a reason of increasing the entropy when constructing a text both in meaning and in structuring level. It is nothing but artificial formatting of lexical units that violate stylistic unity of a phrase. It is marked that one of the important characteristics descending the entropy in the field of emergence is a typical similarity of lexical and semantic resources of the different lingua-cultures in aspects of extra linguistic factors.

Keywords: communicative situation, field of emergence, lingua-culture, entropy

Procedia PDF Downloads 350
617 Peril´s Environment of Energetic Infrastructure Complex System, Modelling by the Crisis Situation Algorithms

Authors: Jiří F. Urbánek, Alena Oulehlová, Hana Malachová, Jiří J. Urbánek Jr.

Abstract:

Crisis situations investigation and modelling are introduced and made within the complex system of energetic critical infrastructure, operating on peril´s environments. Every crisis situations and perils has an origin in the emergency/ crisis event occurrence and they need critical/ crisis interfaces assessment. Here, the emergency events can be expected - then crisis scenarios can be pre-prepared by pertinent organizational crisis management authorities towards their coping; or it may be unexpected - without pre-prepared scenario of event. But the both need operational coping by means of crisis management as well. The operation, forms, characteristics, behaviour and utilization of crisis management have various qualities, depending on real critical infrastructure organization perils, and prevention training processes. An aim is always - better security and continuity of the organization, which successful obtainment needs to find and investigate critical/ crisis zones and functions in critical infrastructure organization models, operating in pertinent perils environment. Our DYVELOP (Dynamic Vector Logistics of Processes) method is disposables for it. Here, it is necessary to derive and create identification algorithm of critical/ crisis interfaces. The locations of critical/ crisis interfaces are the flags of crisis situation in organization of critical infrastructure models. Then, the model of crisis situation will be displayed at real organization of Czech energetic crisis infrastructure subject in real peril environment. These efficient measures are necessary for the infrastructure protection. They will be derived for peril mitigation, crisis situation coping and for environmentally friendly organization survival, continuity and its sustainable development advanced possibilities.

Keywords: algorithms, energetic infrastructure complex system, modelling, peril´s environment

Procedia PDF Downloads 391
616 Exploring Data Stewardship in Fog Networking Using Blockchain Algorithm

Authors: Ruvaitha Banu, Amaladhithyan Krishnamoorthy

Abstract:

IoT networks today solve various consumer problems, from home automation systems to aiding in driving autonomous vehicles with the exploration of multiple devices. For example, in an autonomous vehicle environment, multiple sensors are available on roads to monitor weather and road conditions and interact with each other to aid the vehicle in reaching its destination safely and timely. IoT systems are predominantly dependent on the cloud environment for data storage, and computing needs that result in latency problems. With the advent of Fog networks, some of this storage and computing is pushed to the edge/fog nodes, saving the network bandwidth and reducing the latency proportionally. Managing the data stored in these fog nodes becomes crucial as it might also store sensitive information required for a certain application. Data management in fog nodes is strenuous because Fog networks are dynamic in terms of their availability and hardware capability. It becomes more challenging when the nodes in the network also live a short span, detaching and joining frequently. When an end-user or Fog Node wants to access, read, or write data stored in another Fog Node, then a new protocol becomes necessary to access/manage the data stored in the fog devices as a conventional static way of managing the data doesn’t work in Fog Networks. The proposed solution discusses a protocol that acts by defining sensitivity levels for the data being written and read. Additionally, a distinct data distribution and replication model among the Fog nodes is established to decentralize the access mechanism. In this paper, the proposed model implements stewardship towards the data stored in the Fog node using the application of Reinforcement Learning so that access to the data is determined dynamically based on the requests.

Keywords: IoT, fog networks, data stewardship, dynamic access policy

Procedia PDF Downloads 47
615 Evaluating the Validity of CFD Model of Dispersion in a Complex Urban Geometry Using Two Sets of Experimental Measurements

Authors: Mohammad R. Kavian Nezhad, Carlos F. Lange, Brian A. Fleck

Abstract:

This research presents the validation study of a computational fluid dynamics (CFD) model developed to simulate the scalar dispersion emitted from rooftop sources around the buildings at the University of Alberta North Campus. The ANSYS CFX code was used to perform the numerical simulation of the wind regime and pollutant dispersion by solving the 3D steady Reynolds-averaged Navier-Stokes (RANS) equations on a building-scale high-resolution grid. The validation study was performed in two steps. First, the CFD model performance in 24 cases (eight wind directions and three wind speeds) was evaluated by comparing the predicted flow fields with the available data from the previous measurement campaign designed at the North Campus, using the standard deviation method (SDM), while the estimated results of the numerical model showed maximum average percent errors of approximately 53% and 37% for wind incidents from the North and Northwest, respectively. Good agreement with the measurements was observed for the other six directions, with an average error of less than 30%. In the second step, the reliability of the implemented turbulence model, numerical algorithm, modeling techniques, and the grid generation scheme was further evaluated using the Mock Urban Setting Test (MUST) dispersion dataset. Different statistical measures, including the fractional bias (FB), the geometric mean bias (MG), and the normalized mean square error (NMSE), were used to assess the accuracy of the predicted dispersion field. Our CFD results are in very good agreement with the field measurements.

Keywords: CFD, plume dispersion, complex urban geometry, validation study, wind flow

Procedia PDF Downloads 123
614 Numerical Simulations of Acoustic Imaging in Hydrodynamic Tunnel with Model Adaptation and Boundary Layer Noise Reduction

Authors: Sylvain Amailland, Jean-Hugh Thomas, Charles Pézerat, Romuald Boucheron, Jean-Claude Pascal

Abstract:

The noise requirements for naval and research vessels have seen an increasing demand for quieter ships in order to fulfil current regulations and to reduce the effects on marine life. Hence, new methods dedicated to the characterization of propeller noise, which is the main source of noise in the far-field, are needed. The study of cavitating propellers in closed-section is interesting for analyzing hydrodynamic performance but could involve significant difficulties for hydroacoustic study, especially due to reverberation and boundary layer noise in the tunnel. The aim of this paper is to present a numerical methodology for the identification of hydroacoustic sources on marine propellers using hydrophone arrays in a large hydrodynamic tunnel. The main difficulties are linked to the reverberation of the tunnel and the boundary layer noise that strongly reduce the signal-to-noise ratio. In this paper it is proposed to estimate the reflection coefficients using an inverse method and some reference transfer functions measured in the tunnel. This approach allows to reduce the uncertainties of the propagation model used in the inverse problem. In order to reduce the boundary layer noise, a cleaning algorithm taking advantage of the low rank and sparse structure of the cross-spectrum matrices of the acoustic and the boundary layer noise is presented. This approach allows to recover the acoustic signal even well under the boundary layer noise. The improvement brought by this method is visible on acoustic maps resulting from beamforming and DAMAS algorithms.

Keywords: acoustic imaging, boundary layer noise denoising, inverse problems, model adaptation

Procedia PDF Downloads 320
613 Investigation and Monitoring Method of Vector Density in Kaohsiung City

Authors: Chiu-Wen Chang, I-Yun Chang, Wei-Ting Chen, Hui-Ping Ho, Chao-Ying Pan, Joh-Jong Huang

Abstract:

Dengue is a ‘community disease’ or ‘environmental disease’, as long as the environment exist suitable container (including natural and artificial) for mosquito breeding, once the virus invade will lead to the dengue epidemic. Surveillance of vector density is critical to effective infectious disease control and play an important role in monitoring the dynamics of mosquitoes in community, such as mosquito species, density, distribution area. The objective of this study was to examine the relationship in vector density survey (Breteau index, Adult index, House index, Container index, and Larvae index) form 2014 to 2016 in Kaohsiung City and evaluate the effects of introducing the Breeding Elimination and Appraisal Team (hereinafter referred to as BEAT) as an intervention measure on eliminating dengue vector breeding site started from May 2016. BEAT were performed on people who were suspected of contracting dengue fever, a surrounding area measuring 50 meters by 50 meters was demarcated as the emergency prevention and treatment zone. BEAT would perform weekly vector mosquito inspections and vector mosquito inspections in regions with a high Gravitrap index and assign a risk assessment index to each region. These indices as well as the prevention and treatment results were immediately reported to epidemic prevention-related units every week. The results indicated that, vector indices from 2014 to 2016 showed no statistically significant differences in the Breteau index, adult index, and house index (p > 0.05) but statistically significant differences in the container index and larvae index (p <0.05). After executing the integrated elimination work, container index and larvae index are statistically significant different from 2014 to 2016 in the (p < 0.05). A post hoc test indicated that the container index of 2014 (M = 12.793) was significantly higher than that of 2016 (M = 7.631), and that the larvae index of 2015 (M = 34.065) was significantly lower than that of 2014 (M = 66.867). The results revealed that effective vector density surveillance could highlight the focus breeding site and then implement the immediate control action (BEAT), which successfully decreased the vector density and the risk of dengue epidemic.

Keywords: Breteau index, dengue control, monitoring method, vector density

Procedia PDF Downloads 170
612 Data Collection in Protected Agriculture for Subsequent Big Data Analysis: Methodological Evaluation in Venezuela

Authors: Maria Antonieta Erna Castillo Holly

Abstract:

During the last decade, data analysis, strategic decision making, and the use of artificial intelligence (AI) tools in Latin American agriculture have been a challenge. In some countries, the availability, quality, and reliability of historical data, in addition to the current data recording methodology in the field, makes it difficult to use information systems, complete data analysis, and their support for making the right strategic decisions. This is something essential in Agriculture 4.0. where the increase in the global demand for fresh agricultural products of tropical origin, during all the seasons of the year requires a change in the production model and greater agility in the responses to the consumer market demands of quality, quantity, traceability, and sustainability –that means extensive data-. Having quality information available and updated in real-time on what, how much, how, when, where, at what cost, and the compliance with production quality standards represents the greatest challenge for sustainable and profitable agriculture in the region. The objective of this work is to present a methodological proposal for the collection of georeferenced data from the protected agriculture sector, specifically in production units (UP) with tall structures (Greenhouses), initially for Venezuela, taking the state of Mérida as the geographical framework, and horticultural products as target crops. The document presents some background information and explains the methodology and tools used in the 3 phases of the work: diagnosis, data collection, and analysis. As a result, an evaluation of the process is carried out, relevant data and dashboards are displayed, and the first satellite maps integrated with layers of information in a geographic information system are presented. Finally, some improvement proposals and tentatively recommended applications are added to the process, understanding that their objective is to provide better qualified and traceable georeferenced data for subsequent analysis of the information and more agile and accurate strategic decision making. One of the main points of this study is the lack of quality data treatment in the Latin America area and especially in the Caribbean basin, being one of the most important points how to manage the lack of complete official data. The methodology has been tested with horticultural products, but it can be extended to other tropical crops.

Keywords: greenhouses, protected agriculture, data analysis, geographic information systems, Venezuela

Procedia PDF Downloads 120
611 Exploring the Role of Data Mining in Crime Classification: A Systematic Literature Review

Authors: Faisal Muhibuddin, Ani Dijah Rahajoe

Abstract:

This in-depth exploration, through a systematic literature review, scrutinizes the nuanced role of data mining in the classification of criminal activities. The research focuses on investigating various methodological aspects and recent developments in leveraging data mining techniques to enhance the effectiveness and precision of crime categorization. Commencing with an exposition of the foundational concepts of crime classification and its evolutionary dynamics, this study details the paradigm shift from conventional methods towards approaches supported by data mining, addressing the challenges and complexities inherent in the modern crime landscape. Specifically, the research delves into various data mining techniques, including K-means clustering, Naïve Bayes, K-nearest neighbour, and clustering methods. A comprehensive review of the strengths and limitations of each technique provides insights into their respective contributions to improving crime classification models. The integration of diverse data sources takes centre stage in this research. A detailed analysis explores how the amalgamation of structured data (such as criminal records) and unstructured data (such as social media) can offer a holistic understanding of crime, enriching classification models with more profound insights. Furthermore, the study explores the temporal implications in crime classification, emphasizing the significance of considering temporal factors to comprehend long-term trends and seasonality. The availability of real-time data is also elucidated as a crucial element in enhancing responsiveness and accuracy in crime classification.

Keywords: data mining, classification algorithm, naïve bayes, k-means clustering, k-nearest neigbhor, crime, data analysis, sistematic literature review

Procedia PDF Downloads 56
610 Cercarial Diversity in Freshwater Snails from Selected Freshwater Bodies and Its Implication for Veterinary and Public Health in Kaduna State, Nigeria

Authors: Fatima Muhammad Abdulkadir, D. B. Maikaje, Y. A. Umar

Abstract:

A study conducted to determine cercariae diversity and prevalence of trematode infection in freshwater snails from six freshwater bodies selected by systematic random sampling in Kaduna State was carried from January 2013 to December 2013. Freshwater snails and cercariae harvested from the study sites were morphologically identified. A total of 23,823 freshwater snails were collected from the six freshwater bodies: Bagoma dam, Gimbawa dam, Kangimi dam, Kubacha dam, Manchok water intake and Saminaka water intake. The observed freshwater snail species were: Melanoides tuberculata, Biomphalaria pfeifferi, Bulinus globosus, Lymnaea natalensis, Physa sp., Cleopatra bulimoides, Bellamya unicolor and Lanistes varicus. The freshwater snails were exposed to artificial bright light from a 100 Watt electric bulb in the laboratory to induce cercarial shedding. Of the total freshwater snails collected, 10.55% released one or more types of cercariae. Seven morphological types of cercariae were shed by six freshwater snail species namely: Brevifurcate-apharyngeate distome, Amphistome, Gymnocephalus, Longifurcate-pharyngeate monostome, Longifurcate-pharyngeate distome, Echinostome and Xiphidio cercariae. Infection was monotype in most of the freshwater snails collected; however, Physa species presented a mixed infection with Gymnocephalus and Longifurcate-pharyngeate distome cercariae. B. globosus and B. pfeifferi were the most preferred intermediate hosts with the prevalence of 13.48% and 13.46%, respectively. The diversity and prevalence of cercariae varied among the six freshwater bodies with Manchok water intake having the highest infestation (14.3%) and the least recorded in Kangimi dam (3.9%). There was a correlation trend between the number of freshwater snails and trematode infection with Manchok exhibiting the highest and Bagoma none. The highest cercarial diversity was observed in B. pfeifferi and B. globosus with four morphotypes each, and the lowest was in M. tuberculata with one morphotype. The general distribution of freshwater snails and the trematode cercariae they shed suggests the risk of human and animals to trematodiasis in Manchok community. Public health education to raise awareness on individual and communal action that may control snail breeding sites, prevent transmission and provide access to treatment should be intensified.

Keywords: Cercariae, diversity, freshwater snails, prevalence, trematodiasis

Procedia PDF Downloads 217
609 ZigBee Wireless Sensor Nodes with Hybrid Energy Storage System Based on Li-Ion Battery and Solar Energy Supply

Authors: Chia-Chi Chang, Chuan-Bi Lin, Chia-Min Chan

Abstract:

Most ZigBee sensor networks to date make use of nodes with limited processing, communication, and energy capabilities. Energy consumption is of great importance in wireless sensor applications as their nodes are commonly battery-driven. Once ZigBee nodes are deployed outdoors, limited power may make a sensor network useless before its purpose is complete. At present, there are two strategies for long node and network lifetime. The first strategy is saving energy as much as possible. The energy consumption will be minimized through switching the node from active mode to sleep mode and routing protocol with ultra-low energy consumption. The second strategy is to evaluate the energy consumption of sensor applications as accurately as possible. Erroneous energy model may render a ZigBee sensor network useless before changing batteries. In this paper, we present a ZigBee wireless sensor node with four key modules: a processing and radio unit, an energy harvesting unit, an energy storage unit, and a sensor unit. The processing unit uses CC2530 for controlling the sensor, carrying out routing protocol, and performing wireless communication with other nodes. The harvesting unit uses a 2W solar panel to provide lasting energy for the node. The storage unit consists of a rechargeable 1200 mAh Li-ion battery and a battery charger using a constant-current/constant-voltage algorithm. Our solution to extend node lifetime is implemented. Finally, a long-term sensor network test is used to exhibit the functionality of the solar powered system.

Keywords: ZigBee, Li-ion battery, solar panel, CC2530

Procedia PDF Downloads 367
608 Lamb Waves Wireless Communication in Healthy Plates Using Coherent Demodulation

Authors: Rudy Bahouth, Farouk Benmeddour, Emmanuel Moulin, Jamal Assaad

Abstract:

Guided ultrasonic waves are used in Non-Destructive Testing (NDT) and Structural Health Monitoring (SHM) for inspection and damage detection. Recently, wireless data transmission using ultrasonic waves in solid metallic channels has gained popularity in some industrial applications such as nuclear, aerospace and smart vehicles. The idea is to find a good substitute for electromagnetic waves since they are highly attenuated near metallic components due to Faraday shielding. The proposed solution is to use ultrasonic guided waves such as Lamb waves as an information carrier due to their capability of propagation for long distances. In addition to this, valuable information about the health of the structure could be extracted simultaneously. In this work, the reliable frequency bandwidth for communication is extracted experimentally from dispersion curves at first. Then, an experimental platform for wireless communication using Lamb waves is described and built. After this, coherent demodulation algorithm used in telecommunications is tested for Amplitude Shift Keying, On-Off Keying and Binary Phase Shift Keying modulation techniques. Signal processing parameters such as threshold choice, number of cycles per bit and Bit Rate are optimized. Experimental results are compared based on the average Bit Error Rate. Results have shown high sensitivity to threshold selection for Amplitude Shift Keying and On-Off Keying techniques resulting a Bit Rate decrease. Binary Phase Shift Keying technique shows the highest stability and data rate between all tested modulation techniques.

Keywords: lamb waves communication, wireless communication, coherent demodulation, bit error rate

Procedia PDF Downloads 236
607 HLA-DPB1 Matching on the Outcome of Unrelated Donor Hematopoietic Stem Cell Transplantation

Authors: Shi-xia Xu, Zai-wen Zhang, Ru-xue Chen, Shan Zhou, Xiang-feng Tang

Abstract:

Objective: The clinical influence of HLA-DPB1 mismatches on clinical outcome of HSCT is less clear. This is the first meta-analysis to study the HLA-DPB1 matching statues on clinical outcomes after unrelated donor HSCT. Methods: We searched the CIBMTR, Cochrane Central Register of Controlled Trials (CENTRAL) and related databases (1995.01–2017.06) for all relevant articles. Comparative studies were used to investigate the HLA-DPB1 loci mismatches on clinical outcomes after unrelated donor HSCT, such as the disease-free survival (DFS), overall survival, GVHD, relapse, and transplant-related mortality (TRM). We performed meta-analysis using Review Manager 5.2 software and funnel plot to assess the bias. Results: At first, 1246 articles were retrieved, and 18 studies totaling 26368 patients analyzed. Pooled comparisons of studies found that the HLA-DPB1 mismatched group had a lower rate of DFS than the DPB1-matched group, and lower OS in non-T cell depleted transplantation. The DPB1 mismatched group has a higher incidence of aGVHD and more severe ( ≥ III degree) aGvHD, lower rate of relapse and higher TRM. Moreover, compared with 1-antigen mismatch, 2-antigen mismatched led to a higher risk of TRM and lower relapse rate. Conclusions: This meta-analysis indicated HLA-DPB1 has important influence on survival and transplant-related complications during unrelated donor HSCT and HLA-DPB1 donor selection strategies have been proposed based on a personalized algorithm.

Keywords: human leukocyte antigen, DPB1, transplant, meta-analysis, outcome

Procedia PDF Downloads 285
606 Loading and Unloading Scheduling Problem in a Multiple-Multiple Logistics Network: Modelling and Solving

Authors: Yasin Tadayonrad

Abstract:

Most of the supply chain networks have many nodes starting from the suppliers’ side up to the customers’ side that each node sends/receives the raw materials/products from/to the other nodes. One of the major concerns in this kind of supply chain network is finding the best schedule for loading /unloading the shipments through the whole network by which all the constraints in the source and destination nodes are met and all the shipments are delivered on time. One of the main constraints in this problem is loading/unloading capacity in each source/ destination node at each time slot (e.g., per week/day/hour). Because of the different characteristics of different products/groups of products, the capacity of each node might differ based on each group of products. In most supply chain networks (especially in the Fast-moving consumer goods industry), there are different planners/planning teams working separately in different nodes to determine the loading/unloading timeslots in source/destination nodes to send/receive the shipments. In this paper, a mathematical problem has been proposed to find the best timeslots for loading/unloading the shipments minimizing the overall delays subject to respecting the capacity of loading/unloading of each node, the required delivery date of each shipment (considering the lead-times), and working-days of each node. This model was implemented on python and solved using Python-MIP on a sample data set. Finally, the idea of a heuristic algorithm has been proposed as a way of improving the solution method that helps to implement the model on larger data sets in real business cases, including more nodes and shipments.

Keywords: supply chain management, transportation, multiple-multiple network, timeslots management, mathematical modeling, mixed integer programming

Procedia PDF Downloads 85
605 Screening for Non-hallucinogenic Neuroplastogens as Drug Candidates for the Treatment of Anxiety, Depression, and Posttraumatic Stress Disorder

Authors: Jillian M. Hagel, Joseph E. Tucker, Peter J. Facchini

Abstract:

With the aim of establishing a holistic approach for the treatment of central nervous system (CNS) disorders, we are pursuing a drug development program rapidly progressing through discovery and characterization phases. The drug candidates identified in this program are referred to as neuroplastogens owing to their ability to mediate neuroplasticity, which can be beneficial to patients suffering from anxiety, depression, or posttraumatic stress disorder. These and other related neuropsychiatric conditions are associated with the onset of neuronal atrophy, which is defined as a reduction in the number and/or productivity of neurons. The stimulation of neuroplasticity results in an increase in the connectivity between neurons and promotes the restoration of healthy brain function. We have synthesized a substantial catalogue of proprietary indolethylamine derivatives based on the general structures of serotonin (5-hydroxytryptamine) and psychedelic molecules such as N,N-dimethyltryptamine (DMT) and psilocin (4-hydroxy-DMT) that function as neuroplastogens. A primary objective in our screening protocol is the identification of derivatives associated with a significant reduction in hallucination, which will allow administration of the drug at a dose that induces neuroplasticity and triggers other efficacious outcomes in the treatment of targeted CNS disorders but which does not cause a psychedelic response in the patient. Both neuroplasticity and hallucination are associated with engagement of the 5HT2A receptor, requiring drug candidates differentially coupled to these two outcomes at a molecular level. We use novel and proprietary artificial intelligence algorithms to predict the mode of binding to the 5HT2A receptor, which has been shown to correlate with the hallucinogenic response. Hallucination is tested using the mouse head-twitch response model, whereas mouse marble-burying and sucrose preference assays are used to evaluate anxiolytic and anti-depressive potential. Neuroplasticity is assays using dendritic outgrowth assays and cell-based ELISA analysis. Pharmacokinetics and additional receptor-binding analyses also contribute the selection of lead candidates. A summary of the program is presented.

Keywords: neuroplastogen, non-hallucinogenic, drug development, anxiety, depression, PTSD, indolethylamine derivatives, psychedelic-inspired, 5-HT2A receptor, computational chemistry, head-twitch response behavioural model, neurite outgrowth assay

Procedia PDF Downloads 119
604 A Real-Time Moving Object Detection and Tracking Scheme and Its Implementation for Video Surveillance System

Authors: Mulugeta K. Tefera, Xiaolong Yang, Jian Liu

Abstract:

Detection and tracking of moving objects are very important in many application contexts such as detection and recognition of people, visual surveillance and automatic generation of video effect and so on. However, the task of detecting a real shape of an object in motion becomes tricky due to various challenges like dynamic scene changes, presence of shadow, and illumination variations due to light switch. For such systems, once the moving object is detected, tracking is also a crucial step for those applications that used in military defense, video surveillance, human computer interaction, and medical diagnostics as well as in commercial fields such as video games. In this paper, an object presents in dynamic background is detected using adaptive mixture of Gaussian based analysis of the video sequences. Then the detected moving object is tracked using the region based moving object tracking and inter-frame differential mechanisms to address the partial overlapping and occlusion problems. Firstly, the detection algorithm effectively detects and extracts the moving object target by enhancing and post processing morphological operations. Secondly, the extracted object uses region based moving object tracking and inter-frame difference to improve the tracking speed of real-time moving objects in different video frames. Finally, the plotting method was applied to detect the moving objects effectively and describes the object’s motion being tracked. The experiment has been performed on image sequences acquired both indoor and outdoor environments and one stationary and web camera has been used.

Keywords: background modeling, Gaussian mixture model, inter-frame difference, object detection and tracking, video surveillance

Procedia PDF Downloads 463
603 Predicting Emerging Agricultural Investment Opportunities: The Potential of Structural Evolution Index

Authors: Kwaku Damoah

Abstract:

The agricultural sector is characterized by continuous transformation, driven by factors such as demographic shifts, evolving consumer preferences, climate change, and migration trends. This dynamic environment presents complex challenges for key stakeholders including farmers, governments, and investors, who must navigate these changes to achieve optimal investment returns. To effectively predict market trends and uncover promising investment opportunities, a systematic, data-driven approach is essential. This paper introduces the Structural Evolution Index (SEI), a machine learning-based methodology. SEI is specifically designed to analyse long-term trends and forecast the potential of emerging agricultural products for investment. Versatile in application, it evaluates various agricultural metrics such as production, yield, trade, land use, and consumption, providing a comprehensive view of the evolution within agricultural markets. By harnessing data from the UN Food and Agricultural Organisation (FAOSTAT), this study demonstrates the SEI's capabilities through Comparative Exploratory Analysis and evaluation of international trade in agricultural products, focusing on Malaysia and Singapore. The SEI methodology reveals intricate patterns and transitions within the agricultural sector, enabling stakeholders to strategically identify and capitalize on emerging markets. This predictive framework is a powerful tool for decision-makers, offering crucial insights that help anticipate market shifts and align investments with anticipated returns.

Keywords: agricultural investment, algorithm, comparative exploratory analytics, machine learning, market trends, predictive analytics, structural evolution index

Procedia PDF Downloads 51
602 Anti-Forensic Countermeasure: An Examination and Analysis Extended Procedure for Information Hiding of Android SMS Encryption Applications

Authors: Ariq Bani Hardi

Abstract:

Empowerment of smartphone technology is growing very rapidly in various fields of science. One of the mobile operating systems that dominate the smartphone market today is Android by Google. Unfortunately, the expansion of mobile technology is misused by criminals to hide the information that they store or exchange with each other. It makes law enforcement more difficult to prove crimes committed in the judicial process (anti-forensic). One of technique that used to hide the information is encryption, such as the usages of SMS encryption applications. A Mobile Forensic Examiner or an investigator should prepare a countermeasure technique if he finds such things during the investigation process. This paper will discuss an extension procedure if the investigator found unreadable SMS in android evidence because of encryption. To define the extended procedure, we create and analyzing a dataset of android SMS encryption application. The dataset was grouped by application characteristics related to communication permissions, as well as the availability of source code and the documentation of encryption scheme. Permissions indicate the possibility of how applications exchange the data and keys. Availability of the source code and the encryption scheme documentation can show what the cryptographic algorithm specification is used, how long the key length, how the process of key generation, key exchanges, encryption/decryption is done, and other related information. The output of this paper is an extended or alternative procedure for examination and analysis process of android digital forensic. It can be used to help the investigators while they got a confused cause of SMS encryption during examining and analyzing. What steps should the investigator take, so they still have a chance to discover the encrypted SMS in android evidence?

Keywords: anti-forensic countermeasure, SMS encryption android, examination and analysis, digital forensic

Procedia PDF Downloads 121
601 A Hybrid Energy Storage Module for the Emergency Energy System of the Community Shelter in Yucatán, México

Authors: María Reveles-Miranda, Daniella Pacheco-Catalán

Abstract:

Sierra Papacal commissary is located north of Merida, Yucatan, México, where the indigenous Maya population predominates. Due to its location, the region has an elevation of fewer than 4.5 meters above sea level, with a high risk of flooding associated with storms and hurricanes and a high vulnerability of infrastructure and housing in the presence of strong gusts of wind. In environmental contingencies, the challenge is providing an autonomous electrical supply using renewable energy sources that cover vulnerable populations' health, food, and water pumping needs. To address this challenge, a hybrid energy storage module is proposed for the emergency photovoltaic (PV) system of the community shelter in Sierra Papacal, Yucatán, which combines high-energy-density batteries and high-power-density supercapacitors (SC) in a single module, providing a quick response to energy demand, reducing the thermal stress on batteries and extending their useful life. Incorporating SC in energy storage modules can provide fast response times to power variations and balanced energy extraction, ensuring a more extended period of electrical supply to vulnerable populations during contingencies. The implemented control strategy increases the module's overall performance by ensuring the optimal use of devices and balanced energy exploitation. The operation of the module with the control algorithm is validated with MATLAB/Simulink® and experimental tests.

Keywords: batteries, community shelter, environmental contingencies, hybrid energy storage, isolated photovoltaic system, supercapacitors

Procedia PDF Downloads 71
600 Genetic Differentiation between Members of a Species Complex (Retropinna spp.)

Authors: Md. Rakeb-Ul Islam, Daniel J. Schmidt, Jane M. Hughes

Abstract:

Population connectivity plays an important role in the conservation and recovery of declining species. It affects genetic diversity, adaptive potential and resilience of species in nature. Loss of genetic variation can affect populations by limiting their ability to persist in stressful environmental conditions. Generally, freshwater fishes show higher levels of genetic structuring and subdivision among populations than those inhabiting estuarine or marine environments due to the presence of artificial (e.g. dams) and natural (e.g. mountain ranges) barriers to dispersal in freshwater ecosystems. The Australian smelt (Retropinnidae: Retropinna spp.) is a common freshwater fish species which is widely distributed throughout coastal and inland drainages in South - eastern Australia. These fish are found in a number of habitats from headwaters to lowland sites. They form large shoals in the mid to upper water column and inhabit deep slow – flowing pools as well as shallow fast flowing riffle-runs. Previously, Australian smelt consisted of two described taxa (Retropinna semoni and Retropinna tasmanica), but recently a complex of five or more species has been recognized based on an analysis of allozyme variation. In many area, they spend their entire life cycle within freshwater. Although most populations of the species are thought to be non-diadromous, it is still unclear whether individuals within coastal populations of Australian Retropinna exhibit diadromous migrations or whether fish collected from marine/estuarine environments are vagrants that have strayed out of the freshwater reaches. In this current study, the population structure and genetic differentiation of Australian smelt fish were investigated among eight rivers of South-East Queensland (SEQ), Australia. 11 microsatellite loci were used to examine genetic variation within and among populations. Genetic diversity was very high. Number of alleles ranged from three to twenty. Expected heterozygosity averaged across loci ranged from 0.572 to 0.852. There was a high degree of genetic differentiation among rivers (FST = 0.23), although low levels of genetic differentiation among populations within rivers. These extremely high levels of genetic differentiation suggest that the all smelt in SEQ complete their life history within freshwater, or, if they go to the estuary, they do not migrate to sea. This hypothesis is being tested further with a micro-chemical analysis of their otoliths.

Keywords: diadromous, genetic diversity, microsatellite, otolith

Procedia PDF Downloads 296
599 Iris Recognition Based on the Low Order Norms of Gradient Components

Authors: Iman A. Saad, Loay E. George

Abstract:

Iris pattern is an important biological feature of human body; it becomes very hot topic in both research and practical applications. In this paper, an algorithm is proposed for iris recognition and a simple, efficient and fast method is introduced to extract a set of discriminatory features using first order gradient operator applied on grayscale images. The gradient based features are robust, up to certain extents, against the variations may occur in contrast or brightness of iris image samples; the variations are mostly occur due lightening differences and camera changes. At first, the iris region is located, after that it is remapped to a rectangular area of size 360x60 pixels. Also, a new method is proposed for detecting eyelash and eyelid points; it depends on making image statistical analysis, to mark the eyelash and eyelid as a noise points. In order to cover the features localization (variation), the rectangular iris image is partitioned into N overlapped sub-images (blocks); then from each block a set of different average directional gradient densities values is calculated to be used as texture features vector. The applied gradient operators are taken along the horizontal, vertical and diagonal directions. The low order norms of gradient components were used to establish the feature vector. Euclidean distance based classifier was used as a matching metric for determining the degree of similarity between the features vector extracted from the tested iris image and template features vectors stored in the database. Experimental tests were performed using 2639 iris images from CASIA V4-Interival database, the attained recognition accuracy has reached up to 99.92%.

Keywords: iris recognition, contrast stretching, gradient features, texture features, Euclidean metric

Procedia PDF Downloads 323
598 Assessing the Legacy Effects of Wildfire on Eucalypt Canopy Structure of South Eastern Australia

Authors: Yogendra K. Karna, Lauren T. Bennett

Abstract:

Fire-tolerant eucalypt forests are one of the major forest ecosystems of south-eastern Australia and thought to be highly resistant to frequent high severity wildfires. However, the impact of different severity wildfires on the canopy structure of fire-tolerant forest type is under-studied, and there are significant knowledge gaps in relation to the assessment of tree and stand level canopy structural dynamics and recovery after fire. Assessment of canopy structure is a complex task involving accurate measurements of the horizontal and vertical arrangement of the canopy in space and time. This study examined the utility of multitemporal, small-footprint lidar data to describe the changes in the horizontal and vertical canopy structure of fire-tolerant eucalypt forests seven years after wildfire of different severities from the tree to stand level. Extensive ground measurements were carried out in four severity classes to describe and validate canopy cover and height metrics as they change after wildfire. Several metrics such as crown height and width, crown base height and clumpiness of crown were assessed at tree and stand level using several individual tree top detection and measurement algorithm. Persistent effects of high severity fire 8 years after both on tree crowns and stand canopy were observed. High severity fire increased the crown depth but decreased the crown projective cover leading to more open canopy.

Keywords: canopy gaps, canopy structure, crown architecture, crown projective cover, multi-temporal lidar, wildfire severity

Procedia PDF Downloads 162
597 An Energy-Balanced Clustering Method on Wireless Sensor Networks

Authors: Yu-Ting Tsai, Chiun-Chieh Hsu, Yu-Chun Chu

Abstract:

In recent years, due to the development of wireless network technology, many researchers have devoted to the study of wireless sensor networks. The applications of wireless sensor network mainly use the sensor nodes to collect the required information, and send the information back to the users. Since the sensed area is difficult to reach, there are many restrictions on the design of the sensor nodes, where the most important restriction is the limited energy of sensor nodes. Because of the limited energy, researchers proposed a number of ways to reduce energy consumption and balance the load of sensor nodes in order to increase the network lifetime. In this paper, we proposed the Energy-Balanced Clustering method with Auxiliary Members on Wireless Sensor Networks(EBCAM)based on the cluster routing. The main purpose is to balance the energy consumption on the sensed area and average the distribution of dead nodes in order to avoid excessive energy consumption because of the increasing in transmission distance. In addition, we use the residual energy and average energy consumption of the nodes within the cluster to choose the cluster heads, use the multi hop transmission method to deliver the data, and dynamically adjust the transmission radius according to the load conditions. Finally, we use the auxiliary cluster members to change the delivering path according to the residual energy of the cluster head in order to its load. Finally, we compare the proposed method with the related algorithms via simulated experiments and then analyze the results. It reveals that the proposed method outperforms other algorithms in the numbers of used rounds and the average energy consumption.

Keywords: auxiliary nodes, cluster, load balance, routing algorithm, wireless sensor network

Procedia PDF Downloads 267
596 Evaluation of the Irritation Potential of Three Topical Formulations of Minoxidil 5% + Finasteride 0.1% Using Patch Test

Authors: Joshi Rajiv, Shah Priyank, Thavkar Amit, Rohira Poonam, Mehta Suyog

Abstract:

Topical formulation containing minoxidil and finasteride helps hair growth in the treatment of male androgenetic alopecia. The objective of this study is to compare the irritation potential of three conventional formulations of minoxidil 5% + finasteride 0.1% topical solution of in human patch test. The study was a single centre, double blind, non-randomized controlled study in 53 healthy adult Indian subjects. Occlusive patch test for 24 hours was performed with three formulations of minoxidil 5% + finasteride 0.1% topical solution. Products tested included aqueous based minoxidil 5% + finasteride 0.1% (AnasureTM-F, Sun Pharma, India – Brand A), lipid based minoxidil 5% + finasteride 0.1% (Brand B) and aqueous based minoxidil 5% + finasteride 0.1% (Brand C). Isotonic saline 0.9% and 1% w/w sodium lauryl sulphate were included as negative control and positive control respectively. Patches were applied and removed after 24 hours. The skin reaction was assessed and clinically scored 24 hours after the removal of the patches under constant artificial daylight source using the Draize scale (0-4 points scale for erythema/dryness//wrinkles and for oedema). Follow-up was scheduled after one week to confirm recovery for any reaction. A combined mean score up to 2.0/8.0 indicates a product is “non-irritant” and a score between 2.0/8.0 and 4.0/8.0 indicates “mildly irritant” and a score above 4.0/8.0 indicates “irritant”. The procedure of the patch test followed the principles outlined by the Bureau of Indian Standards (BIS) (IS 4011:2018; Methods of Test for safety evaluation of Cosmetics-3rd revision). Fifty three subjects with mean age 31.9 years (25 males and 28 females) participated in the study. The combined mean score ± standard deviation were: 0.06 ± 0.23 (Brand A), 0.81 ± 0.59 (Brand B), 0.38 ± 0.49 (Brand C), 2.92 ± 0.47 (positive control) and 0.0 ± 0.0 (Negative control). This means the score of Brand A (Sun Pharma product) was significantly lower than that of Brand B (p=0.001) and that of Brand C (p=0.001). The combined mean erythema score ± standard deviation were: 0.06 ± 0.23 (Brand A), 0.81 ± 0.59 (Brand B), 0.38 ± 0.49 (Brand C), 2.09 ± 0.4 (Positive control) and 0.0 ± 0.0 (Negative control). The mean erythema score of Brand A was significantly lower than Brand B (p=0.001) and that of Brand C (p=0.001). Any reaction observed at 24hours after patch removal subsided in a week. All the three topical formulations of minoxidil 5% + finasteride 0.1% were non-irritant. Brand A of minoxidil 5% + finasteride 0.1% (Sun Pharma) was found to be the least irritant than Brand B and Brand C based on the combined mean score and mean erythema score in the human patch test as per the BIS, IS 4011:2018

Keywords: erythema, finasteride, irritation, minoxidil, patch test

Procedia PDF Downloads 70
595 Evaluation of the Irritation Potential of Three Topical Formulations of Minoxidil 2% Using Patch Test

Authors: Sule Pallavi, Shah Priyank, Thavkar Amit, Rohira Poonam, Mehta Suyog

Abstract:

Introduction: Minoxidil has been used topically for a long time to assist hair growth in the management of male androgenetic alopecia. The aim of this study was a comparative assessment of the irritation potential of three commercial formulations of minoxidil 2% topical solution in a human patch test. Methodology: The study was a non-randomized, double-blind, controlled, single-center study of 56 healthy adult Indian subjects. A 24-hour occlusive patch test was conducted with three formulations of minoxidil 2% topical solution. Products tested were aqueous-based minoxidil 2% (AnasureTM 2%, Sun Pharma, India – Brand A), alcohol-based minoxidil 2% (Brand B) and aqueous-based minoxidil 2% (Brand C). Isotonic saline 0.9% and 1% w/w sodium lauryl sulphate as a negative and positive control, respectively, were included. Patches were applied on the back, followed by removal after 24 hours. The Draize scale (0-4 points scale for erythema/dryness/wrinkles and for oedema) was used to evaluate and clinically score the skin reaction under constant artificial daylight 24 hours after the removal of the patches. The patch test was based on the principles outlined by Bureau of Indian Standards (BIS) (IS 4011:2018; Methods of Test for safety evaluation of Cosmetics-3rd revision). A mean combined score up to 2.0/8.0 indicates that a product is “non-irritant,” and a score between 2.0/8.0 and 4.0/8.0 indicates “mildly irritant” and a score above 4.0/8.0 indicates “irritant”. In case of any skin reaction that was observed, a follow-up was planned after one week to confirm recovery. Results: The 56 subjects who participated in the study had a mean age of 28.7 years (28 males and 28 females). The combined mean score ± standard deviation was: 0.09 ± 0.29 (Brand A), 0.29± 0.53 (Brand B), 0.30 ± 0.46 (Brand C), 3.25 ± 0.77 (positive control) and 0.02 ± 0.13 (negative control). This mean score of Brand A (Sun Pharma) was significantly lower than that of Brand B (p=0.016) and that of Brand C (p=0.004). The mean erythema score ± standard deviation was: 0.09 ± 0.29 (Brand A), 0.27 ± 0.49 (Brand B), 0.30 ± 0.46 (Brand C), 2.5 ± 0.66 (positive control) and 0.02 ± 0.13 (negative control). The mean erythema score of Brand A (Sun Pharma) was significantly lower than that of Brand B (p=0.019) and that of Brand C (p=0.004). Reactions that were observed 24 hours after patch removal subsided in a week’s time. Conclusion: Based on the human patch test as per the BIS, IS 4011:2018, all the three topical formulations of minoxidil 2% were found to be non-irritant. Brand A of 2% minoxidil (Sun Pharma) was found to be the least irritant than Brand B and Brand C based on the combined mean score and mean erythema score.

Keywords: erythema, irritation, minoxidil, patch test

Procedia PDF Downloads 74