Search results for: proposed drought severity index
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13144

Search results for: proposed drought severity index

2014 Exploring the Effects of Transcendental Mindfulness Meditation on Anxiety Symptoms in Young Females

Authors: Claudia Cedeno Nadal, Mei-Ling Villafana, Griela Rodriguez, Jessica Martin, Jennifer Martin, Megan Patel

Abstract:

This study systematically examines the impact of Transcendental Mindfulness Meditation on anxiety symptoms in young females aged 18-25. Through a comprehensive literature review, we found consistent evidence supporting the positive influence of Transcendental Mindfulness Meditation on reducing anxiety, enhancing overall well-being, and decreasing perceived stress levels within this demographic. The mechanisms underlying these effects include heightened self-awareness, improved emotional regulation, and the development of effective stress-coping strategies. These findings have significant implications for mental health interventions targeting young females. However, the reviewed studies had some limitations, such as small sample sizes and reliance on self-report measures. To advance this field, future research should focus on larger sample sizes and utilize a broader range of measurement methods, including neuroscience assessments. Additionally, investigating the temporal relationships between Transcendental Mindfulness Meditation, proposed mediators, and anxiety symptoms will help establish causal specificity and a deeper understanding of the precise mechanisms of action. The development of integrative models based on these mechanisms can further enhance the effectiveness of Transcendental Mindfulness Meditation as an intervention for anxiety in this demographic. This study contributes to the current knowledge on the potential benefits of Transcendental Mindfulness Meditation for reducing anxiety in young females, paving the way for more targeted and effective mental health interventions in this population.

Keywords: mindfulness, meditation, anxiety, transcendental mindfulness

Procedia PDF Downloads 70
2013 The Reenactment of Historic Memory and the Ways to Read past Traces through Contemporary Architecture in European Urban Contexts: The Case Study of the Medieval Walls of Naples

Authors: Francesco Scarpati

Abstract:

Because of their long history, ranging from ancient times to the present day, European cities feature many historical layers, whose single identities are represented by traces surviving in the urban design. However, urban transformations, in particular, the ones that have been produced by the property speculation phenomena of the 20th century, often compromised the readability of these traces, resulting in a loss of the historical identities of the single layers. The purpose of this research is, therefore, a reflection on the theme of the reenactment of the historical memory in the stratified European contexts and on how contemporary architecture can help to reveal past signs of the cities. The research work starts from an analysis of a series of emblematic examples that have already provided an original solution to the described problem, going from the architectural detail scale to the urban and landscape scale. The results of these analyses are then applied to the case study of the city of Naples, as an emblematic example of a stratified city, with an ancient Greek origin; a city where it is possible to read most of the traces of its transformations. Particular consideration is given to the trace of the medieval walls of the city, which a long time ago clearly divided the city itself from the outer fields, and that is no longer readable at the current time. Finally, solutions and methods of intervention are proposed to ensure that the trace of the walls, read as a boundary, can be revealed through the contemporary project.

Keywords: contemporary project, historic memory, historic urban contexts, medieval walls, naples, stratified cities, urban traces

Procedia PDF Downloads 264
2012 High School Gain Analytics From National Assessment Program – Literacy and Numeracy and Australian Tertiary Admission Rankin Linkage

Authors: Andrew Laming, John Hattie, Mark Wilson

Abstract:

Nine Queensland Independent high schools provided deidentified student-matched ATAR and NAPLAN data for all 1217 ATAR graduates since 2020 who also sat NAPLAN at the school. Graduating cohorts from the nine schools contained a mean 100 ATAR graduates with previous NAPLAN data from their school. Excluded were vocational students (mean=27) and any ATAR graduates without NAPLAN data (mean=20). Based on Index of Community Socio-Educational Access (ICSEA) prediction, all schools had larger that predicted proportions of their students graduating with ATARs. There were an additional 173 students not releasing their ATARs to their school (14%), requiring this data to be inferred by schools. Gain was established by first converting each student’s strongest NAPLAN domain to a statewide percentile, then subtracting this result from final ATAR. The resulting ‘percentile shift’ was corrected for plausible ATAR participation at each NAPLAN level. Strongest NAPLAN domain had the highest correlation with ATAR (R2=0.58). RESULTS School mean NAPLAN scores fitted ICSEA closely (R2=0.97). Schools achieved a mean cohort gain of two ATAR rankings, but only 66% of students gained. This ranged from 46% of top-NAPLAN decile students gaining, rising to 75% achieving gains outside the top decile. The 54% of top-decile students whose ATAR fell short of prediction lost a mean 4.0 percentiles (or 6.2 percentiles prior to correction for regression to the mean). 71% of students in smaller schools gained, compared to 63% in larger schools. NAPLAN variability in each of the 13 ICSEA1100 cohorts was 17%, with both intra-school and inter-school variation of these values extremely low (0.3% to 1.8%). Mean ATAR change between years in each school was just 1.1 ATAR ranks. This suggests consecutive school cohorts and ICSEA-similar schools share very similar distributions and outcomes over time. Quantile analysis of the NAPLAN/ATAR revealed heteroscedasticity, but splines offered little additional benefit over simple linear regression. The NAPLAN/ATAR R2 was 0.33. DISCUSSION Standardised data like NAPLAN and ATAR offer educators a simple no-cost progression metric to analyse performance in conjunction with their internal test results. Change is expressed in percentiles, or ATAR shift per student, which is layperson intuitive. Findings may also reduce ATAR/vocational stream mismatch, reveal proportions of cohorts meeting or falling short of expectation and demonstrate by how much. Finally, ‘crashed’ ATARs well below expectation are revealed, which schools can reasonably work to minimise. The percentile shift method is neither value-add nor a growth percentile. In the absence of exit NAPLAN testing, this metric is unable to discriminate academic gain from legitimate ATAR-maximizing strategies. But by controlling for ICSEA, ATAR proportion variation and student mobility, it uncovers progression to ATAR metrics which are not currently publicly available. However achieved, ATAR maximisation is a sought-after private good. So long as standardised nationwide data is available, this analysis offers useful analytics for educators and reasonable predictivity when counselling subsequent cohorts about their ATAR prospects.  

Keywords: NAPLAN, ATAR, analytics, measurement, gain, performance, data, percentile, value-added, high school, numeracy, reading comprehension, variability, regression to the mean

Procedia PDF Downloads 68
2011 Machine Learning Facing Behavioral Noise Problem in an Imbalanced Data Using One Side Behavioral Noise Reduction: Application to a Fraud Detection

Authors: Salma El Hajjami, Jamal Malki, Alain Bouju, Mohammed Berrada

Abstract:

With the expansion of machine learning and data mining in the context of Big Data analytics, the common problem that affects data is class imbalance. It refers to an imbalanced distribution of instances belonging to each class. This problem is present in many real world applications such as fraud detection, network intrusion detection, medical diagnostics, etc. In these cases, data instances labeled negatively are significantly more numerous than the instances labeled positively. When this difference is too large, the learning system may face difficulty when tackling this problem, since it is initially designed to work in relatively balanced class distribution scenarios. Another important problem, which usually accompanies these imbalanced data, is the overlapping instances between the two classes. It is commonly referred to as noise or overlapping data. In this article, we propose an approach called: One Side Behavioral Noise Reduction (OSBNR). This approach presents a way to deal with the problem of class imbalance in the presence of a high noise level. OSBNR is based on two steps. Firstly, a cluster analysis is applied to groups similar instances from the minority class into several behavior clusters. Secondly, we select and eliminate the instances of the majority class, considered as behavioral noise, which overlap with behavior clusters of the minority class. The results of experiments carried out on a representative public dataset confirm that the proposed approach is efficient for the treatment of class imbalances in the presence of noise.

Keywords: machine learning, imbalanced data, data mining, big data

Procedia PDF Downloads 130
2010 Optimization of Multi Commodities Consumer Supply Chain: Part 1-Modelling

Authors: Zeinab Haji Abolhasani, Romeo Marian, Lee Luong

Abstract:

This paper and its companions (Part II, Part III) will concentrate on optimizing a class of supply chain problems known as Multi- Commodities Consumer Supply Chain (MCCSC) problem. MCCSC problem belongs to production-distribution (P-D) planning category. It aims to determine facilities location, consumers’ allocation, and facilities configuration to minimize total cost (CT) of the entire network. These facilities can be manufacturer units (MUs), distribution centres (DCs), and retailers/end-users (REs) but not limited to them. To address this problem, three major tasks should be undertaken. At the first place, a mixed integer non-linear programming (MINP) mathematical model is developed. Then, system’s behaviors under different conditions will be observed using a simulation modeling tool. Finally, the most optimum solution (minimum CT) of the system will be obtained using a multi-objective optimization technique. Due to the large size of the problem, and the uncertainties in finding the most optimum solution, integration of modeling and simulation methodologies is proposed followed by developing new approach known as GASG. It is a genetic algorithm on the basis of granular simulation which is the subject of the methodology of this research. In part II, MCCSC is simulated using discrete-event simulation (DES) device within an integrated environment of SimEvents and Simulink of MATLAB® software package followed by a comprehensive case study to examine the given strategy. Also, the effect of genetic operators on the obtained optimal/near optimal solution by the simulation model will be discussed in part III.

Keywords: supply chain, genetic algorithm, optimization, simulation, discrete event system

Procedia PDF Downloads 316
2009 Characterizing Surface Machining-Induced Local Deformation Using Electron Backscatter Diffraction

Authors: Wenqian Zhang, Xuelin Wang, Yujin Hu, Siyang Wang

Abstract:

The subsurface layer of a component plays a significant role in its service performance. Any surface mechanical process during fabrication can introduce a deformed layer near the surface, which can be related to the microstructure alteration and strain hardening, and affects the mechanical properties and corrosion resistance of the material. However, there exists a great difficulty in determining the subsurface deformation induced by surface machining. In this study, electron backscatter diffraction (EBSD) was used to study the deformed layer of surface milled 316 stainless steel. The microstructure change was displayed by the EBSD maps and characterized by misorientation variation. The results revealed that the surface milling resulted in heavily nonuniform deformations in the subsurface layer and even in individual grains. The direction of the predominant grain deformation was about 30-60 deg to the machined surface. Moreover, a local deformation rate (LDR) was proposed to quantitatively evaluate the local deformation degree. Both of the average and maximum LDRs were utilized to characterize the deformation trend along the depth direction. It was revealed that the LDR had a strong correlation with the development of grain and sub-grain boundaries. In this work, a scan step size of 1.2 μm was chosen for the EBSD measurement. A LDR higher than 18 deg/μm indicated a newly developed grain boundary, while a LDR ranged from 2.4 to 18 deg/μm implied the generation of a sub-grain boundary. And a lower LDR than 2.4 deg/μm could only introduce a slighter deformation and no sub-grain boundary was produced. According to the LDR analysis with the evolution of grain or sub grain boundaries, the deformed layer could be classified into four zones: grain broken layer, seriously deformed layer, slightly deformed layer and non-deformed layer.

Keywords: surface machining, EBSD, subsurface layer, local deformation

Procedia PDF Downloads 331
2008 Introduction to Multi-Agent Deep Deterministic Policy Gradient

Authors: Xu Jie

Abstract:

As a key network security method, cryptographic services must fully cope with problems such as the wide variety of cryptographic algorithms, high concurrency requirements, random job crossovers, and instantaneous surges in workloads. Its complexity and dynamics also make it difficult for traditional static security policies to cope with the ever-changing situation. Cyber Threats and Environment. Traditional resource scheduling algorithms are inadequate when facing complex decisionmaking problems in dynamic environments. A network cryptographic resource allocation algorithm based on reinforcement learning is proposed, aiming to optimize task energy consumption, migration cost, and fitness of differentiated services (including user, data, and task security). By modeling the multi-job collaborative cryptographic service scheduling problem as a multiobjective optimized job flow scheduling problem, and using a multi-agent reinforcement learning method, efficient scheduling and optimal configuration of cryptographic service resources are achieved. By introducing reinforcement learning, resource allocation strategies can be adjusted in real time in a dynamic environment, improving resource utilization and achieving load balancing. Experimental results show that this algorithm has significant advantages in path planning length, system delay and network load balancing, and effectively solves the problem of complex resource scheduling in cryptographic services.

Keywords: multi-agent reinforcement learning, non-stationary dynamics, multi-agent systems, cooperative and competitive agents

Procedia PDF Downloads 24
2007 Digital Repository as a Service: Enhancing Access and Preservation of Cultural Heritage Artefacts

Authors: Lefteris Tsipis, Demosthenes Vouyioukas, George Loumos, Antonis Kargas, Dimitris Varoutas

Abstract:

The employment of technology and digitization is crucial for cultural organizations to establish and sustain digital repositories for their cultural heritage artefacts. This utilization is also essential in facilitating the presentation of cultural works and exhibits to a broader audience. Consequently, in this work, we propose a digital repository that functions as Software as a Service (SaaS), primarily promoting the safe storage, display, and sharing of cultural materials, enhancing accessibility, and fostering a deeper understanding and appreciation of cultural heritage. Moreover, the proposed digital repository service is designed as a multitenant architecture, which enables organizations to expand their reach, enhance accessibility, foster collaboration, and ensure the preservation of their content. Specifically, this project aims to assist each cultural institution in organizing its digital cultural assets into collections and feeding other digital platforms, including educational, museum, pedagogical, and games, through appropriate interfaces. Moreover, the creation of this digital repository offers a cutting-edge and effective open-access laboratory solution. It allows organizations to have a significant influence on their audiences by fostering cultural understanding and appreciation. Additionally, it facilitates the connection between different digital repositories and national/European aggregators, promoting collaboration and information sharing. By embracing this solution, cultural institutions can benefit from shared resources and features, such as system updates, backup and recovery services, and data analytics tools, that are provided by the platform.

Keywords: cultural technologies, gaming technologies, web sharing, digital repository

Procedia PDF Downloads 79
2006 An Intelligent Transportation System for Safety and Integrated Management of Railway Crossings

Authors: M. Magrini, D. Moroni, G. Palazzese, G. Pieri, D. Azzarelli, A. Spada, L. Fanucci, O. Salvetti

Abstract:

Railway crossings are complex entities whose optimal management cannot be addressed unless with the help of an intelligent transportation system integrating information both on train and vehicular flows. In this paper, we propose an integrated system named SIMPLE (Railway Safety and Infrastructure for Mobility applied at level crossings) that, while providing unparalleled safety in railway level crossings, collects data on rail and road traffic and provides value-added services to citizens and commuters. Such services include for example alerts, via variable message signs to drivers and suggestions for alternative routes, towards a more sustainable, eco-friendly and efficient urban mobility. To achieve these goals, SIMPLE is organized as a System of Systems (SoS), with a modular architecture whose components range from specially-designed radar sensors for obstacle detection to smart ETSI M2M-compliant camera networks for urban traffic monitoring. Computational unit for performing forecast according to adaptive models of train and vehicular traffic are also included. The proposed system has been tested and validated during an extensive trial held in the mid-sized Italian town of Montecatini, a paradigmatic case where the rail network is inextricably linked with the fabric of the city. Results of the tests are reported and discussed.

Keywords: Intelligent Transportation Systems (ITS), railway, railroad crossing, smart camera networks, radar obstacle detection, real-time traffic optimization, IoT, ETSI M2M, transport safety

Procedia PDF Downloads 497
2005 Effects and Mechanisms of an Online Short-Term Audio-Based Mindfulness Intervention on Wellbeing in Community Settings and How Stress and Negative Affect Influence the Therapy Effects: Parallel Process Latent Growth Curve Modeling of a Randomized Control

Authors: Man Ying Kang, Joshua Kin Man Nan

Abstract:

The prolonged pandemic has posed alarming public health challenges to various parts of the world, and face-to-face mental health treatment is largely discounted for the control of virus transmission, online psychological services and self-help mental health kits have become essential. Online self-help mindfulness-based interventions have proved their effects on fostering mental health for different populations over the globe. This paper was to test the effectiveness of an online short-term audio-based mindfulness (SAM) program in enhancing wellbeing, dispositional mindfulness, and reducing stress and negative affect in community settings in China, and to explore possible mechanisms of how dispositional mindfulness, stress, and negative affect influenced the intervention effects on wellbeing. Community-dwelling adults were recruited via online social networking sites (e.g., QQ, WeChat, and Weibo). Participants (n=100) were randomized into the mindfulness group (n=50) and a waitlist control group (n=50). In the mindfulness group, participants were advised to spend 10–20 minutes listening to the audio content, including mindful-form practices (e.g., eating, sitting, walking, or breathing). Then practice daily mindfulness exercises for 3 weeks (a total of 21 sessions), whereas those in the control group received the same intervention after data collection in the mindfulness group. Participants in the mindfulness group needed to fill in the World Health Organization Five Well-Being Index (WHO), Positive and Negative Affect Schedule (PANAS), Perceived Stress Scale (PSS), and Freiburg Mindfulness Inventory (FMI) four times: at baseline (T0) and at 1 (T1), 2 (T2), and 3 (T3) weeks while those in the waitlist control group only needed to fill in the same scales at pre- and post-interventions. Repeated-measure analysis of variance, paired sample t-test, and independent sample t-test was used to analyze the variable outcomes of the two groups. The parallel process latent growth curve modeling analysis was used to explore the longitudinal moderated mediation effects. The dependent variable was WHO slope from T0 to T3, the independent variable was Group (1=SAM, 2=Control), the mediator was FMI slope from T0 to T3, and the moderator was T0NA and T0PSS separately. The different levels of moderator effects on WHO slope was explored, including low T0NA or T0PSS (Mean-SD), medium T0NA or T0PSS (Mean), and high T0NA or T0PSS (Mean+SD). The results found that SAM significantly improved and predicted higher levels of WHO slope and FMI slope, as well as significantly reduced NA and PSS. FMI slope positively predict WHO slope. FMI slope partially mediated the relationship between SAM and WHO slope. Baseline NA and PSS as the moderators were found to be significant between SAM and WHO slope and between SAM and FMI slope, respectively. The conclusion was that SAM was effective in promoting levels of mental wellbeing, positive affect, and dispositional mindfulness as well as reducing negative affect and stress in community settings in China. SAM improved wellbeing faster through the faster enhancement of dispositional mindfulness. Participants with medium-to-high negative affect and stress buffered the therapy effects of SAM on wellbeing improvement speed.

Keywords: mindfulness, negative affect, stress, wellbeing, randomized control trial

Procedia PDF Downloads 109
2004 Effects of Cooking and Drying on the Phenolic Compounds, and Antioxidant Activity of Cleome gynandra (Spider Plant)

Authors: E. Kayitesi, S. Moyo, V. Mavumengwana

Abstract:

Cleome gynandra (spider plant) is an African green leafy vegetable categorized as an indigenous, underutilized and has been reported to contain essential phenolic compounds. Phenolic compounds play a significant role in human diets due to their proposed health benefits. These compounds however may be affected by different processing methods such as cooking and drying. Cleome gynandra was subjected to boiling, steam blanching, and drying processes and analysed for Total Phenolic Content (TPC), Total Flavonoid Content (TFC), antioxidant activity and flavonoid composition. Cooking and drying significantly (p < 0.05) increased the levels of phenolic compounds and antioxidant activity of the vegetable. The boiled sample filtrate exhibited the lowest TPC followed by the raw sample while the steamed sample depicted the highest TPC levels. Antioxidant activity results showed that steamed sample showed the highest DPPH, FRAP and ABTS with mean values of 499.38 ± 2.44, 578.68 ± 5.19, and 214.39 ± 12.33 μM Trolox Equivalent/g respectively. An increase in quercetin-3-rutinoside, quercetin-rhamnoside and kaempferol-3-rutinoside occurred after all the cooking and drying methods employed. Cooking and drying exerted positive effects on the vegetable’s phenolic content, antioxidant activity as a whole, but with varied effects on the individual flavonoid molecules. The results obtained help in defining the importance of African green leafy vegetable and resultant processed products as functional foods and their potential to exert health promoting properties.

Keywords: Cleome gynandra, phenolic compounds, cooking, drying, health promoting properties

Procedia PDF Downloads 170
2003 A Sharp Interface Model for Simulating Seawater Intrusion in the Coastal Aquifer of Wadi Nador (Algeria)

Authors: Abdelkader Hachemi, Boualem Remini

Abstract:

Seawater intrusion is a significant challenge faced by coastal aquifers in the Mediterranean basin. This study aims to determine the position of the sharp interface between seawater and freshwater in the aquifer of Wadi Nador, located in the Wilaya of Tipaza, Algeria. A numerical areal sharp interface model using the finite element method is developed to investigate the spatial and temporal behavior of seawater intrusion. The aquifer is assumed to be homogeneous and isotropic. The simulation results are compared with geophysical prospection data obtained through electrical methods in 2011 to validate the model. The simulation results demonstrate a good agreement with the geophysical prospection data, confirming the accuracy of the sharp interface model. The position of the sharp interface in the aquifer is found to be approximately 1617 meters from the sea. Two scenarios are proposed to predict the interface position for the year 2024: one without pumping and the other with pumping. The results indicate a noticeable retreat of the sharp interface position in the first scenario, while a slight decline is observed in the second scenario. The findings of this study provide valuable insights into the dynamics of seawater intrusion in the Wadi Nador aquifer. The predicted changes in the sharp interface position highlight the potential impact of pumping activities on the aquifer's vulnerability to seawater intrusion. This study emphasizes the importance of implementing measures to manage and mitigate seawater intrusion in coastal aquifers. The sharp interface model developed in this research can serve as a valuable tool for assessing and monitoring the vulnerability of aquifers to seawater intrusion.

Keywords: seawater intrusion, sharp interface, coastal aquifer, algeria

Procedia PDF Downloads 120
2002 Electrochemical Synthesis of Copper Nanoparticles

Authors: Juan Patricio Ibáñez, Exequiel López

Abstract:

A method for synthesizing copper nanoparticles through an electrochemical approach is proposed, employing surfactants to stabilize the size of the newly formed nanoparticles. The electrolyte was made up of a matrix of H₂SO₄ (190 g/L) having Cu²⁺ (from 3.2 to 9.5 g/L), sodium dodecyl sulfate -SDS- (from 0.5 to 1.0 g/L) and Tween 80 (from 0 to 7.5 mL/L). Tween 80 was used in a molar relation of 1 to 1 with SDS. A glass cell was used, which was in a thermostatic water bath to keep the system temperature, and the electrodes were cathodic copper as an anode and stainless steel 316-L as a cathode. This process was influenced by the control exerted through the initial copper concentration in the electrolyte and the applied current density. Copper nanoparticles of electrolytic purity, exhibiting a spherical morphology of varying sizes with low dispersion, were successfully produced, contingent upon the chemical composition of the electrolyte and current density. The minimum size achieved was 3.0 nm ± 0.9 nm, with an average standard deviation of 2.2 nm throughout the entire process. The deposited copper mass ranged from 0.394 g to 1.848 g per hour (over an area of 25 cm²), accompanied by an average Faradaic efficiency of 30.8% and an average specific energy consumption of 4.4 kWh/kg. The chemical analysis of the product employed X-ray powder diffraction (XRD), while physical characteristics such as size and morphology were assessed using atomic force microscopy (AFM). It was identified that the initial concentration of copper and the current density are the variables defining the size and dispersion of the nanoparticles, as they serve as reactants in the cathodic half-reaction. The presence of surfactants stabilizes the nanoparticle size as their molecules adsorb onto the nanoparticle surface, forming a thick barrier that prevents mass transfer with the exterior and halts further growth.

Keywords: copper nanopowder, electrochemical synthesis, current density, surfactant stabilizer

Procedia PDF Downloads 63
2001 Chemical, Biochemical and Sensory Evaluation of a Quadrimix Complementary Food Developed from Sorghum, Groundnut, Crayfish and Pawpaw Blends

Authors: Ogechi Nzeagwu, Assumpta Osuagwu, Charlse Nkwoala

Abstract:

Malnutrition in infants due to poverty, poor feeding practices, and high cost of commercial complementary foods among others is a concern in developing countries. The study evaluated the proximate, vitamin and mineral compositions, antinutrients and functional properties, biochemical, haematological and sensory evaluation of complementary food made from sorghum, groundnut, crayfish and paw-paw flour blends using standard procedures. The blends were formulated on protein requirement of infants (18 g/day) using Nutrisurvey linear programming software in ratio of sorghum(S), groundnut(G), crayfish(C) and pawpaw(P) flours as 50:25:10:15(SGCP1), 60:20:10:10 (SGCP2), 60:15:15:10 (SGCP3) and 60:10:20:10 (SGCP4). Plain-pap (fermented maize flour)(TCF) and cerelac (commercial complementary food) served as basal and control diets. Thirty weanling male albino rats aged 28-35 days weighing 33-60 g were purchased and used for the study. The rats after acclimatization were fed with gruel produced with the experimental diets and the control with water ad libitum daily for 35days. Effect of the blends on lipid profile, blood glucose, haematological (RBC, HB, PCV, MCV), liver and kidney function and weight gain of the rats were assessed. Acceptability of the gruel was conducted at the end of rat feeding on forty mothers of infants’ ≥ 6 months who gave their informed consent to participate using a 9 point hedonic scale. Data was analyzed for means and standard deviation, analysis of variance and means were separated using Duncan multiple range test and significance judged at 0.05, all using SPSS version 22.0. The results indicated that crude protein, fibre, ash and carbohydrate of the formulated diets were either comparable or higher than values in cerelac. The formulated diets (SGCP1- SGCP4) were significantly (P>0.05) higher in vitamin A and thiamin compared to cerelac. The iron content of the formulated diets SGCP1- SGCP4 (4.23-6.36 mg/100) were within the recommended iron intake of infants (0.55 mg/day). Phytate (1.56-2.55 mg/100g) and oxalate (0.23-0.35 mg/100g) contents of the formulated diets were within the permissible limits of 0-5%. In functional properties, bulk density, swelling index, % dispersibility and water absorption capacity significantly (P<0.05) increased and compared favourably with cerelac. The essential amino acids of the formulated blends were within the amino acid profile of the FAO/WHO/UNU reference protein for children 0.5 -2 years of age. Urea concentration of rats fed with SGCP1-SGCP4 (19.48 mmol/L),(23.76 mmol/L),(24.07 mmol/L),(23.65 mmol/L) respectively was significantly higher than that of rat fed cerelac (16.98 mmol/L); however, plain pap had the least value (9.15 mmol/L). Rats fed with SGCP1-SGCP4 (116 mg/dl), (119 mg/dl), (115 mg/dl), (117 mg/dl) respectively had significantly higher glucose levels those fed with cerelac (108 mg/dl). Liver function parameters (AST, ALP and ALT), lipid profile (triglyceride, HDL, LDL, VLDL) and hematological parameters of rats fed with formulated diets were within normal range. Rats fed SGCP1 gained more weight (90.45 g) than other rats fed with SGCP2-SGCP4 (71.65 g, 79.76 g, 75.68 g), TCF (20.13 g) and cerelac (59.06 g). In all the sensory attributes, the control was preferred with respect to the formulated diets. The formulated diets were generally adequate and may likely have potentials to meet nutrient requirements of infants as complementary food.

Keywords: biochemical, chemical evaluation, complementary food, quadrimix

Procedia PDF Downloads 170
2000 Audit on Compliance with Ottawa Ankle Rules in Ankle Radiograph Requests

Authors: Daud Muhammad

Abstract:

Introduction: Ankle radiographs are frequently requested in Emergency Departments (ED) for patients presenting with traumatic ankle pain. The Ottawa Ankle Rules (OAR) serve as a clinical guideline to determine the necessity of these radiographs, aiming to reduce unnecessary imaging. This audit was conducted to evaluate the adequacy of clinical information provided in radiograph requests in relation to the OAR. Methods: A retrospective analysis was performed on 50 consecutive ankle radiograph requests under ED clinicians' names for patients aged above 5 years, specifically excluding follow-up radiographs for known fractures. The study assessed whether the provided clinical information met the criteria outlined by the OAR. Results: The audit revealed that none of the 50 radiograph requests contained sufficient information to satisfy the Ottawa Ankle Rules. Furthermore, 10 out of the 50 radiographs (20%) identified fractures. Discussion: The findings indicate a significant lack of adherence to the OAR, suggesting potential overuse of radiography and unnecessary patient exposure to radiation. This non-compliance may also contribute to increased healthcare costs and resource utilization, as well as possible delays in diagnosis and treatment. Recommendations: To address these issues, the following recommendations are proposed: (1) Education and Training: Enhance awareness and training among ED clinicians regarding the OAR. (2) Standardised Request Forms: Implement changes to imaging request forms to mandate relevant information according to the OAR. (3) Scan Vetting: Promote awareness among radiographers to discuss the appropriateness of scan requests with clinicians. (4) Regular re-audits should be conducted to monitor improvements in compliance.

Keywords: Ottawa ankle rules, ankle radiographs, emergency department, traumatic pain

Procedia PDF Downloads 45
1999 Biophysical Consideration in the Interaction of Biological Cell Membranes with Virus Nanofilaments

Authors: Samaneh Farokhirad, Fatemeh Ahmadpoor

Abstract:

Biological membranes are constantly in contact with various filamentous soft nanostructures that either reside on their surface or are being transported between the cell and its environment. In particular, viral infections are determined by the interaction of viruses (such as filovirus) with cell membranes, membrane protein organization (such as cytoskeletal proteins and actin filament bundles) has been proposed to influence the mechanical properties of lipid membranes, and the adhesion of filamentous nanoparticles influence their delivery yield into target cells or tissues. The goal of this research is to integrate the rapidly increasing but still fragmented experimental observations on the adhesion and self-assembly of nanofilaments (including filoviruses, actin filaments, as well as natural and synthetic nanofilaments) on cell membranes into a general, rigorous, and unified knowledge framework. The global outbreak of the coronavirus disease in 2020, which has persisted for over three years, highlights the crucial role that nanofilamentbased delivery systems play in human health. This work will unravel the role of a unique property of all cell membranes, namely flexoelectricity, and the significance of nanofilaments’ flexibility in the adhesion and self-assembly of nanofilaments on cell membranes. This will be achieved utilizing a set of continuum mechanics, statistical mechanics, and molecular dynamics and Monte Carlo simulations. The findings will help address the societal needs to understand biophysical principles that govern the attachment of filoviruses and flexible nanofilaments onto the living cells and provide guidance on the development of nanofilament-based vaccines for a range of diseases, including infectious diseases and cancer.

Keywords: virus nanofilaments, cell mechanics, computational biophysics, statistical mechanics

Procedia PDF Downloads 94
1998 Simplified Stress Gradient Method for Stress-Intensity Factor Determination

Authors: Jeries J. Abou-Hanna

Abstract:

Several techniques exist for determining stress-intensity factors in linear elastic fracture mechanics analysis. These techniques are based on analytical, numerical, and empirical approaches that have been well documented in literature and engineering handbooks. However, not all techniques share the same merit. In addition to overly-conservative results, the numerical methods that require extensive computational effort, and those requiring copious user parameters hinder practicing engineers from efficiently evaluating stress-intensity factors. This paper investigates the prospects of reducing the complexity and required variables to determine stress-intensity factors through the utilization of the stress gradient and a weighting function. The heart of this work resides in the understanding that fracture emanating from stress concentration locations cannot be explained by a single maximum stress value approach, but requires use of a critical volume in which the crack exists. In order to understand the effectiveness of this technique, this study investigated components of different notch geometry and varying levels of stress gradients. Two forms of weighting functions were employed to determine stress-intensity factors and results were compared to analytical exact methods. The results indicated that the “exponential” weighting function was superior to the “absolute” weighting function. An error band +/- 10% was met for cases ranging from a steep stress gradient in a sharp v-notch to the less severe stress transitions of a large circular notch. The incorporation of the proposed method has shown to be a worthwhile consideration.

Keywords: fracture mechanics, finite element method, stress intensity factor, stress gradient

Procedia PDF Downloads 135
1997 Feature Selection of Personal Authentication Based on EEG Signal for K-Means Cluster Analysis Using Silhouettes Score

Authors: Jianfeng Hu

Abstract:

Personal authentication based on electroencephalography (EEG) signals is one of the important field for the biometric technology. More and more researchers have used EEG signals as data source for biometric. However, there are some disadvantages for biometrics based on EEG signals. The proposed method employs entropy measures for feature extraction from EEG signals. Four type of entropies measures, sample entropy (SE), fuzzy entropy (FE), approximate entropy (AE) and spectral entropy (PE), were deployed as feature set. In a silhouettes calculation, the distance from each data point in a cluster to all another point within the same cluster and to all other data points in the closest cluster are determined. Thus silhouettes provide a measure of how well a data point was classified when it was assigned to a cluster and the separation between them. This feature renders silhouettes potentially well suited for assessing cluster quality in personal authentication methods. In this study, “silhouettes scores” was used for assessing the cluster quality of k-means clustering algorithm is well suited for comparing the performance of each EEG dataset. The main goals of this study are: (1) to represent each target as a tuple of multiple feature sets, (2) to assign a suitable measure to each feature set, (3) to combine different feature sets, (4) to determine the optimal feature weighting. Using precision/recall evaluations, the effectiveness of feature weighting in clustering was analyzed. EEG data from 22 subjects were collected. Results showed that: (1) It is possible to use fewer electrodes (3-4) for personal authentication. (2) There was the difference between each electrode for personal authentication (p<0.01). (3) There is no significant difference for authentication performance among feature sets (except feature PE). Conclusion: The combination of k-means clustering algorithm and silhouette approach proved to be an accurate method for personal authentication based on EEG signals.

Keywords: personal authentication, K-mean clustering, electroencephalogram, EEG, silhouettes

Procedia PDF Downloads 285
1996 Culture of Primary Cortical Neurons on Hydrophobic Nanofibers Induces the Formation of Organoid-Like Structures

Authors: Nick Weir, Robert Stevens, Alan Hargreaves, Martin McGinnity, Chris Tinsley

Abstract:

Hydrophobic materials have previously demonstrated the ability to elevate cell-cell interactions and promote the formation of neural networks whilst aligned nanofibers demonstrate the ability to induce extensive neurite outgrowth in an aligned manner. Hydrophobic materials typically elicit an immune response upon implantation and thus materials used for implantation are typically hydrophilic. Poly-L-lactic acid (PLLA) is a hydrophobic, non-immunogenic, FDA approved material that can be electrospun to form aligned nanofibers. Primary rat cortical neurons cultured for 10 days on aligned PLLA nanofibers formed 3D cell clusters, approximately 800 microns in diameter. Neurites that extended from these clusters were highly aligned due to the alignment of the nanofibers they were cultured upon and fasciculation was also evident. Plasma treatment of the PLLA nanofibers prior to seeding of cells significantly reduced the hydrophobicity and abolished the cluster formation and neurite fasciculation, whilst reducing the extent and directionality of neurite outgrowth; it is proposed that hydrophobicity induces the changes to cellular behaviors. Aligned PLLA nanofibers induced the formation of a structure that mimics the grey-white matter compartmentalization that is observed in vivo and thus represents a step forward in generating organoids or biomaterial-based implants. Upon implantation into the brain, the biomaterial architectures described here may provide a useful platform for both brain repair and brain remodeling initiatives.

Keywords: hydrophobicity, nanofibers, neurite fasciculation, neurite outgrowth, PLLA

Procedia PDF Downloads 160
1995 Optimization Technique for the Contractor’s Portfolio in the Bidding Process

Authors: Taha Anjamrooz, Sareh Rajabi, Salwa Bheiry

Abstract:

Selection between the available projects in bidding processes for the contractor is one of the essential areas to concentrate on. It is important for the contractor to choose the right projects within its portfolio during the tendering stage based on certain criteria. It should align the bidding process with its origination strategies and goals as a screening process to have the right portfolio pool to start with. Secondly, it should set the proper framework and use a suitable technique in order to optimize its selection process for concertation purpose and higher efforts during the tender stage with goals of success and winning. In this research paper, a two steps framework proposed to increase the efficiency of the contractor’s bidding process and the winning chance of getting the new projects awarded. In this framework, initially, all the projects pass through the first stage screening process, in which the portfolio basket will be evaluated and adjusted in accordance with the organization strategies to the reduced version of the portfolio pool, which is in line with organization activities. In the second stage, the contractor uses linear programming to optimize the portfolio pool based on available resources such as manpower, light equipment, heavy equipment, financial capability, return on investment, and success rate of winning the bid. Therefore, this optimization model will assist the contractor in utilizing its internal resource to its maximum and increase its winning chance for the new project considering past experience with clients, built-relation between two parties, and complexity in the exertion of the projects. The objective of this research will be to increase the contractor's winning chance in the bidding process based on the success rate and expected return on investment.

Keywords: bidding process, internal resources, optimization, contracting portfolio management

Procedia PDF Downloads 142
1994 Real-Time Generative Architecture for Mesh and Texture

Authors: Xi Liu, Fan Yuan

Abstract:

In the evolving landscape of physics-based machine learning (PBML), particularly within fluid dynamics and its applications in electromechanical engineering, robot vision, and robot learning, achieving precision and alignment with researchers' specific needs presents a formidable challenge. In response, this work proposes a methodology that integrates neural transformation with a modified smoothed particle hydrodynamics model for generating transformed 3D fluid simulations. This approach is useful for nanoscale science, where the unique and complex behaviors of viscoelastic medium demand accurate neurally-transformed simulations for materials understanding and manipulation. In electromechanical engineering, the method enhances the design and functionality of fluid-operated systems, particularly microfluidic devices, contributing to advancements in nanomaterial design, drug delivery systems, and more. The proposed approach also aligns with the principles of PBML, offering advantages such as multi-fluid stylization and consistent particle attribute transfer. This capability is valuable in various fields where the interaction of multiple fluid components is significant. Moreover, the application of neurally-transformed hydrodynamical models extends to manufacturing processes, such as the production of microelectromechanical systems, enhancing efficiency and cost-effectiveness. The system's ability to perform neural transfer on 3D fluid scenes using a deep learning algorithm alongside physical models further adds a layer of flexibility, allowing researchers to tailor simulations to specific needs across scientific and engineering disciplines.

Keywords: physics-based machine learning, robot vision, robot learning, hydrodynamics

Procedia PDF Downloads 66
1993 Environmental Impacts of Point and Non-Point Source Pollution in Krishnagiri Reservoir: A Case Study in South India

Authors: N. K. Ambujam, V. Sudha

Abstract:

Reservoirs are being contaminated all around the world with point source and Non-Point Source (NPS) pollution. The most common NPS pollutants are sediments and nutrients. Krishnagiri Reservoir (KR) has been chosen for the present case study, which is located in the tropical semi-arid climatic zone of Tamil Nadu, South India. It is the main source of surface water in Krishnagiri district to meet the freshwater demands. The reservoir has lost about 40% of its water holding capacity due to sedimentation over the period of 50 years. Hence, from the research and management perspective, there is a need for a sound knowledge on the spatial and seasonal variations of KR water quality. The present study encompasses the specific objectives as (i) to investigate the longitudinal heterogeneity and seasonal variations of physicochemical parameters, nutrients and biological characteristics of KR water and (ii) to examine the extent of degradation of water quality in KR. 15 sampling points were identified by uniform stratified method and a systematic monthly sampling strategy was selected due to high dynamic nature in its hydrological characteristics. The physicochemical parameters, major ions, nutrients and Chlorophyll a (Chl a) were analysed. Trophic status of KR was classified by using Carlson's Trophic State Index (TSI). All statistical analyses were performed by using Statistical Package for Social Sciences programme, version-16.0. Spatial maps were prepared for Chl a using Arc GIS. Observations in KR pointed out that electrical conductivity and major ions are highly variable factors as it receives inflow from the catchment with different land use activities. The study of major ions in KR exhibited different trends in their values and it could be concluded that as the monsoon progresses the major ions in the water decreases or water quality stabilizes. The inflow point of KR showed comparatively higher concentration of nutrients including nitrate, soluble reactive phosphorus (SRP), total phosphors (TP), total suspended phosphorus (TSP) and total dissolved phosphorus (TDP) during monsoon seasons. This evidently showed the input of significant amount of nutrients from the catchment side through agricultural runoff. High concentration of TDP and TSP at the lacustrine zone of the reservoir during summer season evidently revealed that there was a significant release of phosphorus from the bottom sediments. Carlson’s TSI of KR ranged between 81 and 92 during northeast monsoon and summer seasons. High and permanent Cyanobacterial bloom in KR could be mainly due to the internal loading of phosphorus from the bottom sediments. According to Carlson’s TSI classification Krishnagiri reservoir was ranked in the hyper-eutrophic category. This study provides necessary basic data on the spatio-temporal variations of water quality in KR and also proves the impact of point and NPS pollution from the catchment area. High TSI warrants a greater threat for the recovery of internal P loading and hyper-eutrophic condition of KR. Several expensive internal measures for the reduction of internal loading of P were introduced by many scientists. However, the outcome of the present research suggests for the innovative algae harvesting technique for the removal of sediment nutrients.

Keywords: NPS pollution, nutrients, hyper-eutrophication, krishnagiri reservoir

Procedia PDF Downloads 324
1992 Data and Biological Sharing Platforms in Community Health Programs: Partnership with Rural Clinical School, University of New South Wales and Public Health Foundation of India

Authors: Vivian Isaac, A. T. Joteeshwaran, Craig McLachlan

Abstract:

The University of New South Wales (UNSW) Rural Clinical School has a strategic collaborative focus on chronic disease and public health. Our objectives are to understand rural environmental and biological interactions in vulnerable community populations. The UNSW Rural Clinical School translational model is a spoke and hub network. This spoke and hub model connects rural data and biological specimens with city based collaborative public health research networks. Similar spoke and hub models are prevalent across research centers in India. The Australia-India Council grant was awarded so we could establish sustainable public health and community research collaborations. As part of the collaborative network we are developing strategies around data and biological sharing platforms between Indian Institute of Public Health, Public Health Foundation of India (PHFI), Hyderabad and Rural Clinical School UNSW. The key objective is to understand how research collaborations are conducted in India and also how data can shared and tracked with external collaborators such as ourselves. A framework to improve data sharing for research collaborations, including DNA was proposed as a project outcome. The complexities of sharing biological data has been investigated via a visit to India. A flagship sustainable project between Rural Clinical School UNSW and PHFI would illustrate a model of data sharing platforms.

Keywords: data sharing, collaboration, public health research, chronic disease

Procedia PDF Downloads 450
1991 A Review of Process Safety Management for Small and Medium Business in Malaysia

Authors: Muhammad Afiq Anaqi Bin Baharudin, Muhammad Izamuddin Bin Mohd Nasir, Syarifuddin Bin Sujuanda, Muhammad Syahmi Rusyaidi Bin Sham Suddin, Danish Hakimi Bin kamaruzaman, Muhammad Haqimi Nazim Bin Hasmanizam, Mohammad Akmal Zakwan Bin Amran, Muhammad Alparizi Bin Latif

Abstract:

In particular, for small and medium enterprises (SMBs) in Malaysia, process safety management (PSM) is a crucial component of industrial safety. Limited resources, a lack of technical know-how, and linguistic and cultural obstacles are just a few of the difficulties SMBs in Malaysia encounter while putting PSM programmes into practice. A number of studies have emphasised the significance of leadership commitment, hazard identification and assessment, and employee involvement in the execution of effective PSM programmes, which are crucial for preventing accidents and incidents. In the literature, there has been a lot of discussion on the creation of specialised PSM frameworks for SMBs in Malaysia. Several studies have proposed implementation frameworks for PSM programmes that are based on recognised worldwide standards. Despite the significance of PSM in ensuring industrial safety, there are still a number of gaps in the literature on PSM in Malaysian SMBs. These gaps include the need for additional research on the efficiency of PSM programmes in reducing accidents and incidents in SMBs as well as the development of more specialised approaches to implementing PSM programmes in SMBs with limited resources and technical expertise. The goal of this review is to give a thorough overview of the body of research on PSM in Malaysian SMBs while highlighting important findings, points of contention, and knowledge gaps that need to be filled in.

Keywords: process safety management, occupational safety and health (OSH), small businesses, medium businesses, malaysia

Procedia PDF Downloads 126
1990 Optimization of Smart Beta Allocation by Momentum Exposure

Authors: J. B. Frisch, D. Evandiloff, P. Martin, N. Ouizille, F. Pires

Abstract:

Smart Beta strategies intend to be an asset management revolution with reference to classical cap-weighted indices. Indeed, these strategies allow a better control on portfolios risk factors and an optimized asset allocation by taking into account specific risks or wishes to generate alpha by outperforming indices called 'Beta'. Among many strategies independently used, this paper focuses on four of them: Minimum Variance Portfolio, Equal Risk Contribution Portfolio, Maximum Diversification Portfolio, and Equal-Weighted Portfolio. Their efficiency has been proven under constraints like momentum or market phenomenon, suggesting a reconsideration of cap-weighting.
 To further increase strategy return efficiency, it is proposed here to compare their strengths and weaknesses inside time intervals corresponding to specific identifiable market phases, in order to define adapted strategies depending on pre-specified situations. 
Results are presented as performance curves from different combinations compared to a benchmark. If a combination outperforms the applicable benchmark in well-defined actual market conditions, it will be preferred. It is mainly shown that such investment 'rules', based on both historical data and evolution of Smart Beta strategies, and implemented according to available specific market data, are providing very interesting optimal results with higher return performance and lower risk.
 Such combinations have not been fully exploited yet and justify present approach aimed at identifying relevant elements characterizing them.

Keywords: smart beta, minimum variance portfolio, equal risk contribution portfolio, maximum diversification portfolio, equal weighted portfolio, combinations

Procedia PDF Downloads 340
1989 Voltage Stabilization of Hybrid PV and Battery Systems by Considering Temperature and Irradiance Changes in Standalone Operation

Authors: S. Jalilzadeh, S. M. Mohseni Bonab

Abstract:

Solar and battery energy storage systems are very useful for consumers who live in deprived areas and do not have access to electricity distribution networks. Nowadays one of the problems that photo voltaic systems (PV) have changing of output power in temperature and irradiance variations, which directly affects the load that is connected to photo voltaic systems. In this paper, with considering the fact that the solar array varies with change in temperature and solar power radiation, a voltage stabilizer system of a load connected to photo voltaic array is designed to stabilize the load voltage and to transfer surplus power of the battery. Also, in proposed hybrid system, the needed load power amount is supplemented considering the voltage stabilization in standalone operation for supplying unbalanced AC load. Electrical energy storage system for voltage control and improvement of the performance of PV by a DC/DC converter is connected to the DC bus. The load is also feed by an AC/DC converter. In this paper, when the voltage increases in its reference limit, the battery gets charged by the photo voltaic array and when it decreases in its defined limit, the power gets injected to the DC bus by this battery. The constant of DC bus Voltage is the cause for the reduced harmonics generated by the inverter. In addition, a series of filters are provided in the inverter output in to reduced harmonics. The inverter control circuit is designed that the voltage and frequency of the load remain almost constant at different load conditions. This paper has focused on controlling strategies of converters to improve their performance.

Keywords: photovoltaic array (PV), DC/DC Boost converter, battery converter, inverters control

Procedia PDF Downloads 485
1988 Evaluation of Synthesis and Structure Elucidation of Some Benzimidazoles as Antimicrobial Agents

Authors: Ozlem Temiz Arpaci, Meryem Tasci, Hakan Goker

Abstract:

Benzimidazole, a structural isostere of indol and purine nuclei that can interact with biopolymers, can be identified as master key. So that benzimidazole compounds are important fragments in medicinal chemistry because of their wide range of biological activities including antimicrobial activity. We planned to synthesize some benzimidazole compounds for developing new antimicrobial drug candidates. In this study, we put some heterocyclic rings on second position and an amidine group on the fifth position of benzimidazole ring and synthesized them using a multiple step procedure. For the synthesis of the compounds, as the first step, 4-chloro-3-nitrobenzonitrile was reacted with cyclohexylamine in dimethyl formamide. Imidate esters (compound 2) were then prepared with absolute ethanol saturated with dry HCl gas. These imidate esters which were not too stable were converted to compound 3 by passing ammonia gas through ethanol. At the Pd / C catalyst, the nitro group is reduced to the amine group (compound 4). Finally, various aldehyde derivatives were reacted with sodium metabisulfite addition products to give compound 5-20. Melting points were determined on a Buchi B-540 melting point apparatus in open capillary tubes and are uncorrected. Elemental analyses were done a Leco CHNS 932 elemental analyzer. 1H-NMR and 13C-NMR spectra were recorded on a Varian Mercury 400 MHz spectrometer using DMSO-d6. Mass spectra were acquired on a Waters Micromass ZQ using the ESI(+) method. The structures of them were supported by spectral data. The 1H-NMR, 13C NMR and mass spectra and elemental analysis results agree with those of the proposed structures. Antimicrobial activity studies of the synthesized compounds are under the investigation.

Keywords: benzimidazoles, synthesis, structure elucidation, antimicrobial

Procedia PDF Downloads 155
1987 Chaotic Electronic System with Lambda Diode

Authors: George Mahalu

Abstract:

The Chua diode has been configured over time in various ways, using electronic structures like operational amplifiers (AOs) or devices with gas or semiconductors. When discussing the use of semiconductor devices, tunnel diodes (Esaki diodes) are most often considered, and more recently, transistorized configurations such as lambda diodes. The paperwork proposed here uses in the modeling a lambda diode type configuration consisting of two junction field effect transistors (JFET). The original scheme is created in the MULTISIM electronic simulation environment and is analyzed in order to identify the conditions for the appearance of evolutionary unpredictability specific to nonlinear dynamic systems with chaos-induced behavior. The chaotic deterministic oscillator is one autonomous type, a fact that places it in the class of Chua’s type oscillators, the only significant and most important difference being the presence of a nonlinear device like the one mentioned structure above. The chaotic behavior is identified both by means of strange attractor-type trajectories and visible during the simulation and by highlighting the hypersensitivity of the system to small variations of one of the input parameters. The results obtained through simulation and the conclusions drawn are useful in the further research of ways to implement such constructive electronic solutions in theoretical and practical applications related to modern small signal amplification structures, to systems for encoding and decoding messages through various modern ways of communication, as well as new structures that can be imagined both in modern neural networks and in those for the physical implementation of some requirements imposed by current research with the aim of obtaining practically usable solutions in quantum computing and quantum computers.

Keywords: chua, diode, memristor, chaos

Procedia PDF Downloads 88
1986 Realizing the Full Potential of Islamic Banking System: Proposed Suitable Legal Framework for Islamic Banking System in Tanzania

Authors: Maulana Ayoub Ali, Pradeep Kulshrestha

Abstract:

Laws of any given secular state have a huge contribution in the growth of the Islamic banking system because the system uses conventional laws to govern its activities. Therefore, the former should be ready to accommodate the latter in order to make the Islamic banking system work properly without affecting the current conventional banking system and therefore without affecting its system. Islamic financial rules have been practiced since the birth of Islam. Following the recent world economic challenges in the financial sector, a quick rebirth of the contemporary Islamic ethical banking system took place. The coming of the Islamic banking system is due to various reasons including but not limited to the failure of the interest based economy in solving financial problems around the globe. Therefore, the Islamic banking system has been adopted as an alternative banking system in order to recover the highly damaged global financial sector. But the Islamic banking system has been facing a number of challenges which hinder its smooth operation in different parts of the world. It has not been the aim of this paper to discuss other challenges rather than the legal ones, but the same was partly discussed when it was justified that it was proper to do so. Generally, there are so many things which have been discovered in the course of writing this paper. The most important part is the issue of the regulatory and supervisory framework for the Islamic banking system in Tanzania and in other nations is considered to be a crucial part for the development of the Islamic banking industry. This paper analyses what has been observed in the study on that area and recommends for necessary actions to be taken on board in a bid to make Islamic banking system reach its climax of serving the larger community by providing ethical, equitable, affordable, interest-free and society cantered banking system around the globe.

Keywords: Islamic banking, interest free banking, ethical banking, legal framework

Procedia PDF Downloads 149
1985 Reversible Information Hitting in Encrypted JPEG Bitstream by LSB Based on Inherent Algorithm

Authors: Vaibhav Barve

Abstract:

Reversible information hiding has drawn a lot of interest as of late. Being reversible, we can restore unique computerized data totally. It is a plan where mystery data is put away in digital media like image, video, audio to maintain a strategic distance from unapproved access and security reason. By and large JPEG bit stream is utilized to store this key data, first JPEG bit stream is encrypted into all around sorted out structure and then this secret information or key data is implanted into this encrypted region by marginally changing the JPEG bit stream. Valuable pixels suitable for information implanting are computed and as indicated by this key subtle elements are implanted. In our proposed framework we are utilizing RC4 algorithm for encrypting JPEG bit stream. Encryption key is acknowledged by framework user which, likewise, will be used at the time of decryption. We are executing enhanced least significant bit supplanting steganography by utilizing genetic algorithm. At first, the quantity of bits that must be installed in a guaranteed coefficient is versatile. By utilizing proper parameters, we can get high capacity while ensuring high security. We are utilizing logistic map for shuffling of bits and utilization GA (Genetic Algorithm) to find right parameters for the logistic map. Information embedding key is utilized at the time of information embedding. By utilizing precise picture encryption and information embedding key, the beneficiary can, without much of a stretch, concentrate the incorporated secure data and totally recoup the first picture and also the original secret information. At the point when the embedding key is truant, the first picture can be recouped pretty nearly with sufficient quality without getting the embedding key of interest.

Keywords: data embedding, decryption, encryption, reversible data hiding, steganography

Procedia PDF Downloads 288