Search results for: independent sensors
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3460

Search results for: independent sensors

280 Preparation of Papers - Developing a Leukemia Diagnostic System Based on Hybrid Deep Learning Architectures in Actual Clinical Environments

Authors: Skyler Kim

Abstract:

An early diagnosis of leukemia has always been a challenge to doctors and hematologists. On a worldwide basis, it was reported that there were approximately 350,000 new cases in 2012, and diagnosing leukemia was time-consuming and inefficient because of an endemic shortage of flow cytometry equipment in current clinical practice. As the number of medical diagnosis tools increased and a large volume of high-quality data was produced, there was an urgent need for more advanced data analysis methods. One of these methods was the AI approach. This approach has become a major trend in recent years, and several research groups have been working on developing these diagnostic models. However, designing and implementing a leukemia diagnostic system in real clinical environments based on a deep learning approach with larger sets remains complex. Leukemia is a major hematological malignancy that results in mortality and morbidity throughout different ages. We decided to select acute lymphocytic leukemia to develop our diagnostic system since acute lymphocytic leukemia is the most common type of leukemia, accounting for 74% of all children diagnosed with leukemia. The results from this development work can be applied to all other types of leukemia. To develop our model, the Kaggle dataset was used, which consists of 15135 total images, 8491 of these are images of abnormal cells, and 5398 images are normal. In this paper, we design and implement a leukemia diagnostic system in a real clinical environment based on deep learning approaches with larger sets. The proposed diagnostic system has the function of detecting and classifying leukemia. Different from other AI approaches, we explore hybrid architectures to improve the current performance. First, we developed two independent convolutional neural network models: VGG19 and ResNet50. Then, using both VGG19 and ResNet50, we developed a hybrid deep learning architecture employing transfer learning techniques to extract features from each input image. In our approach, fusing the features from specific abstraction layers can be deemed as auxiliary features and lead to further improvement of the classification accuracy. In this approach, features extracted from the lower levels are combined into higher dimension feature maps to help improve the discriminative capability of intermediate features and also overcome the problem of network gradient vanishing or exploding. By comparing VGG19 and ResNet50 and the proposed hybrid model, we concluded that the hybrid model had a significant advantage in accuracy. The detailed results of each model’s performance and their pros and cons will be presented in the conference.

Keywords: acute lymphoblastic leukemia, hybrid model, leukemia diagnostic system, machine learning

Procedia PDF Downloads 169
279 From Clients to Colleagues: Supporting the Professional Development of Survivor Social Work Students

Authors: Stephanie Jo Marchese

Abstract:

This oral presentation is a reflective piece regarding current social work teaching methods that value and devalue the lived experiences of survivor students. This presentation grounds the term ‘survivor’ in feminist frameworks. A survivor-defined approach to feminist advocacy assumes an individual’s agency, considers each case and needs independent of generalizations, and provides resources and support to empower victims. Feminist ideologies are ripe arenas to update and influence the rapport-building schools of social work have with these students. Survivor-based frameworks are rooted in nuanced understandings of intersectional realities, staunchly combat both conscious and unconscious deficit lenses wielded against victims, elevate lived experiences to the realm of experiential expertise, and offer alternatives to traditional power structures and knowledge exchanges. Actively importing a survivor framework into the methodology of social work teaching breaks open barriers many survivor students have faced in institutional settings, this author included. The profession of social work is at an important crux of change, both in the United States and globally. The United States is currently undergoing a radical change in its citizenry and outlier communities have taken to the streets again in opposition to their othered-ness. New waves of students are entering this field, emboldened by their survival of personal and systemic oppressions- heavily influenced by third-wave feminism, critical race theory, queer theory, among other post-structuralist ideologies. Traditional models of sociological and psychological studies are actively being challenged. The profession of social work was not founded on the diagnosis of disorders but rather a grassroots-level activism that heralded and demanded resources for oppressed communities. Institutional and classroom acceptance and celebration of survivor narratives can catapult the resurgence of these values needed in the profession’s service-delivery models and put social workers back in the driver's seat of social change (a combined advocacy and policy perspective), moving away from outsider-based intervention models. Survivor students should be viewed as agents of change, not solely former victims and clients. The ideas of this presentation proposal are supported through various qualitative interviews, as well as reviews of ‘best practices’ in the field of education that incorporate feminist methods of inclusion and empowerment. Curriculum and policy recommendations are also offered.

Keywords: deficit lens bias, empowerment theory, feminist praxis, inclusive teaching models, strengths-based approaches, social work teaching methods

Procedia PDF Downloads 270
278 Assumption of Cognitive Goals in Science Learning

Authors: Mihail Calalb

Abstract:

The aim of this research is to identify ways for achieving sustainable conceptual understanding within science lessons. For this purpose, a set of teaching and learning strategies, parts of the theory of visible teaching and learning (VTL), is studied. As a result, a new didactic approach named "learning by being" is proposed and its correlation with educational paradigms existing nowadays in science teaching domain is analysed. In the context of VTL the author describes the main strategies of "learning by being" such as guided self-scaffolding, structuring of information, and recurrent use of previous knowledge or help seeking. Due to the synergy effect of these learning strategies applied simultaneously in class, the impact factor of learning by being on cognitive achievement of students is up to 93 % (the benchmark level is equal to 40% when an experienced teacher applies permanently the same conventional strategy during two academic years). The key idea in "learning by being" is the assumption by the student of cognitive goals. From this perspective, the article discusses the role of student’s personal learning effort within several teaching strategies employed in VTL. The research results emphasize that three mandatory student – related moments are present in each constructivist teaching approach: a) students’ personal learning effort, b) student – teacher mutual feedback and c) metacognition. Thus, a successful educational strategy will target to achieve an involvement degree of students into the class process as high as possible in order to make them not only know the learning objectives but also to assume them. In this way, we come to the ownership of cognitive goals or students’ deep intrinsic motivation. A series of approaches are inherent to the students’ ownership of cognitive goals: independent research (with an impact factor on cognitive achievement equal to 83% according to the results of VTL); knowledge of success criteria (impact factor – 113%); ability to reveal similarities and patterns (impact factor – 132%). Although it is generally accepted that the school is a public service, nonetheless it does not belong to entertainment industry and in most of cases the education declared as student – centered actually hides the central role of the teacher. Even if there is a proliferation of constructivist concepts, mainly at the level of science education research, we have to underline that conventional or frontal teaching, would never disappear. Research results show that no modern method can replace an experienced teacher with strong pedagogical content knowledge. Such a teacher will inspire and motivate his/her students to love and learn physics. The teacher is precisely the condensation point for an efficient didactic strategy – be it constructivist or conventional. In this way, we could speak about "hybridized teaching" where both the student and the teacher have their share of responsibility. In conclusion, the core of "learning by being" approach is guided learning effort that corresponds to the notion of teacher–student harmonic oscillator, when both things – guidance from teacher and student’s effort – are equally important.

Keywords: conceptual understanding, learning by being, ownership of cognitive goals, science learning

Procedia PDF Downloads 152
277 Optimization of Metal Pile Foundations for Solar Power Stations Using Cone Penetration Test Data

Authors: Adrian Priceputu, Elena Mihaela Stan

Abstract:

Our research addresses a critical challenge in renewable energy: improving efficiency and reducing the costs associated with the installation of ground-mounted photovoltaic (PV) panels. The most commonly used foundation solution is metal piles - with various sections adapted to soil conditions and the structural model of the panels. However, direct foundation systems are also sometimes used, especially in brownfield sites. Although metal micropiles are generally the first design option, understanding and predicting their bearing capacity, particularly under varied soil conditions, remains an open research topic. CPT Method and Current Challenges: Metal piles are favored for PV panel foundations due to their adaptability, but existing design methods rely heavily on costly and time-consuming in situ tests. The Cone Penetration Test (CPT) offers a more efficient alternative by providing valuable data on soil strength, stratification, and other key characteristics with reduced resources. During the test, a cone-shaped probe is pushed into the ground at a constant rate. Sensors within the probe measure the resistance of the soil to penetration, divided into cone penetration resistance and shaft friction resistance. Despite some existing CPT-based design approaches for metal piles, these methods are often cumbersome and difficult to apply. They vary significantly due to soil type and foundation method, and traditional approaches like the LCPC method involve complex calculations and extensive empirical data. The method was developed by testing 197 piles on a wide range of ground conditions, but the tested piles were very different from the ones used for PV pile foundations, making the method less accurate and practical for steel micropiles. Project Objectives and Methodology: Our research aims to develop a calculation method for metal micropile foundations using CPT data, simplifying the complex relationships involved. The goal is to estimate the pullout bearing capacity of piles without additional laboratory tests, streamlining the design process. To achieve this, a case study was selected which will serve for the development of an 80ha solar power station. Four testing locations were chosen spread throughout the site. At each location, two types of steel profiles (H160 and C100) were embedded into the ground at various depths (1.5m and 2.0m). The piles were tested for pullout capacity under natural and inundated soil conditions. CPT tests conducted nearby served as calibration points. The results served for the development of a preliminary equation for estimating pullout capacity. Future Work: The next phase involves validating and refining the proposed equation on additional sites by comparing CPT-based forecasts with in situ pullout tests. This validation will enhance the accuracy and reliability of the method, potentially transforming the foundation design process for PV panels.

Keywords: cone penetration test, foundation optimization, solar power stations, steel pile foundations

Procedia PDF Downloads 18
276 Family Carers' Experiences in Striving for Medical Care and Finding Their Solutions for Family Members with Mental Illnesses

Authors: Yu-Yu Wang, Shih-Hua Hsieh, Ru-Shian Hsieh

Abstract:

Wishes and choices being respected, and the right to be supported rather than coerced, have been internationally recognized as the human rights of persons with mental illness. In Taiwan, ‘coerced hospitalization’ has become difficult since the revision of the mental health legislation in 2007. Despite trend towards human rights, the real problem families face when their family members are in mental health crisis is the lack of alternative services. This study aims to explore: 1) When is hospitalization seen as the only solution by family members? 2) What are the barriers for arranging hospitalization, and how are they managed? 3) What have family carers learned, in their experiences of caring for their family members with mental illness? To answer these questions, qualitative approach was adopted, and focus group interviews were taken to collect data. This study includes 24 family carers. The main findings of this research include: First, hospital is the last resort for carers in helplessness. Family carers tend to do everything they could to provide care at home for their family members with mental illness. Carers seek hospitalization only when a patient’s behavior is too violent, weird, and/or abnormal, and beyond their ability to manage. Hospitalization, nevertheless, is never an easy choice. Obstacles emanate from the attitudes of the medical doctors, the restricted areas of ambulance service, and insufficient information from the carers’ part. On the other hand, with some professionals’ proactive assistance, access to medical care while in crisis becomes possible. Some family carers obtained help from the medical doctor, nurse, therapist and social workers. Some experienced good help from policemen, taxi drivers, and security guards at the hospital. The difficulty in accessing medical care prompts carers to work harder on assisting their family members with mental illness to stay in stable states. Carers found different ways of helping the ‘person’ to get along with the ‘illness’ and have better quality of life. Taking back ‘the right to control’ in utilizing medication, from passiveness to negotiating with medical doctors and seeking alternative therapies, are seen in many carers’ efforts. Besides, trying to maintain regular activities in daily life and play normal family roles are also experienced as important. Furthermore, talking with the patient as a person is also important. The authors conclude that in order to protect the human rights of persons with mental illness, it is crucial to make the medical care system more flexible and to make the services more humane: sufficient information should be provided and communicated, and efforts should be made to maintain the person’s social roles and to support the family.

Keywords: family carers, independent living, mental health crisis, persons with mental illness

Procedia PDF Downloads 283
275 Persistent Ribosomal In-Frame Mis-Translation of Stop Codons as Amino Acids in Multiple Open Reading Frames of a Human Long Non-Coding RNA

Authors: Leonard Lipovich, Pattaraporn Thepsuwan, Anton-Scott Goustin, Juan Cai, Donghong Ju, James B. Brown

Abstract:

Two-thirds of human genes do not encode any known proteins. Aside from long non-coding RNA (lncRNA) genes with recently-discovered functions, the ~40,000 non-protein-coding human genes remain poorly understood, and a role for their transcripts as de-facto unconventional messenger RNAs has not been formally excluded. Ribosome profiling (Riboseq) predicts translational potential, but without independent evidence of proteins from lncRNA open reading frames (ORFs), ribosome binding of lncRNAs does not prove translation. Previously, we mass-spectrometrically documented translation of specific lncRNAs in human K562 and GM12878 cells. We now examined lncRNA translation in human MCF7 cells, integrating strand-specific Illumina RNAseq, Riboseq, and deep mass spectrometry in biological quadruplicates performed at two core facilities (BGI, China; City of Hope, USA). We excluded known-protein matches. UCSC Genome Browser-assisted manual annotation of imperfect (tryptic-digest-peptides)-to-(lncRNA-three-frame-translations) alignments revealed three peptides hypothetically explicable by 'stop-to-nonstop' in-frame replacement of stop codons by amino acids in two ORFs of the lncRNA MMP24-AS1. To search for this phenomenon genomewide, we designed and implemented a novel pipeline, matching tryptic-digest spectra to wildcard-instead-of-stop versions of repeat-masked, six-frame, whole-genome translations. Along with singleton putative stop-to-nonstop events affecting four other lncRNAs, we identified 24 additional peptides with stop-to-nonstop in-frame substitutions from multiple positive-strand MMP24-AS1 ORFs. Only UAG and UGA, never UAA, stop codons were impacted. All MMP24-AS1-matching spectra met the same significance thresholds as high-confidence known-protein signatures. Targeted resequencing of MMP24-AS1 genomic DNA and cDNA from the same samples did not reveal any mutations, polymorphisms, or sequencing-detectable RNA editing. This unprecedented apparent gene-specific violation of the genetic code highlights the importance of matching peptides to whole-genome, not known-genes-only, ORFs in mass-spectrometry workflows, and suggests a new mechanism enhancing the combinatorial complexity of the proteome. Funding: NIH Director’s New Innovator Award 1DP2-CA196375 to LL.

Keywords: genetic code, lncRNA, long non-coding RNA, mass spectrometry, proteogenomics, ribo-seq, ribosome, RNAseq

Procedia PDF Downloads 211
274 Healthcare Associated Infections in an Intensive Care Unit in Tunisia: Incidence and Risk Factors

Authors: Nabiha Bouafia, Asma Ben Cheikh, Asma Ammar, Olfa Ezzi, Mohamed Mahjoub, Khaoula Meddeb, Imed Chouchene, Hamadi Boussarsar, Mansour Njah

Abstract:

Background: Hospital acquired infections (HAI) cause significant morbidity, mortality, length of stay and hospital costs, especially in the intensive care unit (ICU), because of the debilitated immune systems of their patients and exposure to invasive devices. The aims of this study were to determine the rate and the risk factors of HAI in an ICU of a university hospital in Tunisia. Materials/Methods: A prospective study was conducted in the 8-bed adult medical ICU of a University Hospital (Sousse Tunisia) during 14 months from September 15th, 2015 to November 15th, 2016. Patients admitted for more than 48h were included. Their surveillance was stopped after the discharge from ICU or death. HAIs were defined according to standard Centers for Disease Control and Prevention criteria. Risk factors were analyzed by conditional stepwise logistic regression. The p-value of < 0.05 was considered significant. Results: During the study, 192 patients had admitted for more than 48 hours. Their mean age was 59.3± 18.20 years and 57.1% were male. Acute respiratory failure was the main reason of admission (72%). The mean SAPS II score calculated at admission was 32.5 ± 14 (range: 6 - 78). The exposure to the mechanical ventilation (MV) and the central venous catheter were observed in 169 (88 %) and 144 (75 %) patients, respectively. Seventy-three patients (38.02%) developed 94 HAIs. The incidence density of HAIs was 41.53 per 1000 patient day. Mortality rate in patients with HAIs was 65.8 %( n= 48). Regarding the type of infection, Ventilator Associated Pneumoniae (VAP) and central venous catheter Associated Infections (CVC AI) were the most frequent with Incidence density: 14.88/1000 days of MV for VAP and 20.02/1000 CVC days for CVC AI. There were 5 Peripheral Venous Catheter Associated Infections, 2 urinary tract infections, and 21 other HAIs. Gram-negative bacteria were the most common germs identified in HAIs: Multidrug resistant Acinetobacter Baumanii (45%) and Klebsiella pneumoniae (10.96%) were the most frequently isolated. Univariate analysis showed that transfer from another hospital department (p= 0.001), intubation (p < 10-4), tracheostomy (p < 10-4), age (p=0.028), grade of acute respiratory failure (p=0.01), duration of sedation (p < 10-4), number of CVC (p < 10-4), length of mechanical ventilation (p < 10-4) and length of stay (p < 10-4), were associated to high risk of HAIS in ICU. Multivariate analysis reveals that independent risk factors for HAIs are: transfer from another hospital department: OR=13.44, IC 95% [3.9, 44.2], p < 10-4, duration of sedation: OR= 1.18, IC 95% [1.049, 1.325], p=0.006, high number of CVC: OR=2.78, IC 95% [1.73, 4.487], p < 10-4, and length of stay in ICU: OR= 1.14, IC 95% [1.066,1.22], p < 10-4. Conclusion: Prevention of nosocomial infections in ICUs is a priority of health care systems all around the world. Yet, their control requires an understanding of epidemiological data collected in these units.

Keywords: healthcare associated infections, incidence, intensive care unit, risk factors

Procedia PDF Downloads 353
273 Mass Flux and Forensic Assessment: Informed Remediation Decision Making at One of Canada’s Most Polluted Sites

Authors: Tony R. Walker, N. Devin MacAskill, Andrew Thalhiemer

Abstract:

Sydney Harbour, Nova Scotia, Canada has long been subject to effluent and atmospheric inputs of contaminants, including thousands of tons of PAHs from a large coking and steel plant which operated in Sydney for nearly a century. Contaminants comprised of coal tar residues which were discharged from coking ovens into a small tidal tributary, which became known as the Sydney Tar Ponds (STPs), and subsequently discharged into Sydney Harbour. An Environmental Impact Statement concluded that mobilization of contaminated sediments posed unacceptable ecological risks, therefore immobilizing contaminants in the STPs using solidification and stabilization was identified as a primary source control remediation option to mitigate against continued transport of contaminated sediments from the STPs into Sydney Harbour. Recent developments in contaminant mass flux techniques focus on understanding “mobile” vs. “immobile” contaminants at remediation sites. Forensic source evaluations are also increasingly used for understanding origins of PAH contaminants in soils or sediments. Flux and forensic source evaluation-informed remediation decision-making uses this information to develop remediation end point goals aimed at reducing off-site exposure and managing potential ecological risk. This study included reviews of previous flux studies, calculating current mass flux estimates and a forensic assessment using PAH fingerprint techniques, during remediation of one of Canada’s most polluted sites at the STPs. Historically, the STPs was thought to be the major source of PAH contamination in Sydney Harbour with estimated discharges of nearly 800 kg/year of PAHs. However, during three years of remediation monitoring only 17-97 kg/year of PAHs were discharged from the STPs, which was also corroborated by an independent PAH flux study during the first year of remediation which estimated 119 kg/year. The estimated mass efflux of PAHs from the STPs during remediation was in stark contrast to ~2000 kg loading thought necessary to cause a short term increase in harbour sediment PAH concentrations. These mass flux estimates during remediation were also between three to eight times lower than PAHs discharged from the STPs a decade prior to remediation, when at the same time, government studies demonstrated on-going reduction in PAH concentrations in harbour sediments. Flux results were also corroborated using forensic source evaluations using PAH fingerprint techniques which found a common source of PAHs for urban soils, marine and aquatic sediments in and around Sydney. Coal combustion (from historical coking) and coal dust transshipment (from current coal transshipment facilities), are likely the principal source of PAHs in these media and not migration of PAH laden sediments from the STPs during a large scale remediation project.

Keywords: contaminated sediment, mass flux, forensic source evaluations, remediation

Procedia PDF Downloads 221
272 Simulation Research of Innovative Ignition System of ASz62IR Radial Aircraft Engine

Authors: Miroslaw Wendeker, Piotr Kacejko, Mariusz Duk, Pawel Karpinski

Abstract:

The research in the field of aircraft internal combustion engines is currently driven by the needs of decreasing fuel consumption and CO2 emissions, while fulfilling the level of safety. Currently, reciprocating aircraft engines are found in sports, emergency, agricultural and recreation aviation. Technically, they are most at a pre-war knowledge of the theory of operation, design and manufacturing technology, especially if compared to that high level of development of automotive engines. Typically, these engines are driven by carburetors of a quite primitive construction. At present, due to environmental requirements and dealing with a climate change, it is beneficial to develop aircraft piston engines and adopt the achievements of automotive engineering such as computer-controlled low-pressure injection, electronic ignition control and biofuels. The paper describes simulation research of the innovative power and control systems for the aircraft radial engine of high power. Installing an electronic ignition system in the radial aircraft engine is a fundamental innovative idea of this solution. Consequently, the required level of safety and better functionality as compared to the today’s plug system can be guaranteed. In this framework, this research work focuses on describing a methodology for optimizing the electronically controlled ignition system. This attempt can reduce emissions of toxic compounds as a result of lowered fuel consumption, optimized combustion and engine capability of efficient combustion of ecological fuels. New, redundant elements of the control system can improve the safety of aircraft. Consequently, the required level of safety and better functionality as compared to the today’s plug system can be guaranteed. The simulation research aimed to determine the vulnerability of the values measured (they were planned as the quantities measured by the measurement systems) to determining the optimal ignition angle (the angle of maximum torque at a given operating point). The described results covered: a) research in steady states; b) velocity ranging from 1500 to 2200 rpm (every 100 rpm); c) loading ranging from propeller power to maximum power; d) altitude ranging according to the International Standard Atmosphere from 0 to 8000 m (every 1000 m); e) fuel: automotive gasoline ES95. The three models of different types of ignition coil (different energy discharge) were studied. The analysis aimed at the optimization of the design of the innovative ignition system for an aircraft engine. The optimization involved: a) the optimization of the measurement systems; b) the optimization of actuator systems. The studies enabled the research on the vulnerability of the signals to the control of the ignition timing. Accordingly, the number and type of sensors were determined for the ignition system to achieve its optimal performance. The results confirmed the limited benefits, in terms of fuel consumption. Thus, including spark management in the optimization is mandatory to significantly decrease the fuel consumption. This work has been financed by the Polish National Centre for Research and Development, INNOLOT, under Grant Agreement No. INNOLOT/I/1/NCBR/2013.

Keywords: piston engine, radial engine, ignition system, CFD model, engine optimization

Procedia PDF Downloads 363
271 Vortex Flows under Effects of Buoyant-Thermocapillary Convection

Authors: Malika Imoula, Rachid Saci, Renee Gatignol

Abstract:

A numerical investigation is carried out to analyze vortex flows in a free surface cylinder, driven by the independent rotation and differentially heated boundaries. As a basic uncontrolled isothermal flow, we consider configurations which exhibit steady axisymmetric toroidal type vortices which occur at the free surface; under given rates of the bottom disk uniform rotation and for selected aspect ratios of the enclosure. In the isothermal case, we show that sidewall differential rotation constitutes an effective kinematic means of flow control: the reverse flow regions may be suppressed under very weak co-rotation rates, while an enhancement of the vortex patterns is remarked under weak counter-rotation. However, in this latter case, high rates of counter-rotation reduce considerably the strength of the meridian flow and cause its confinement to a narrow layer on the bottom disk, while the remaining bulk flow is diffusion dominated and controlled by the sidewall rotation. The main control parameters in this case are the rotational Reynolds number, the cavity aspect ratio and the rotation rate ratio defined. Then, the study proceeded to consider the sensitivity of the vortex pattern, within the Boussinesq approximation, to a small temperature gradient set between the ambient fluid and an axial thin rod mounted on the cavity axis. Two additional parameters are introduced; namely, the Richardson number Ri and the Marangoni number Ma (or the thermocapillary Reynolds number). Results revealed that reducing the rod length induces the formation of on-axis bubbles instead of toroidal structures. Besides, the stagnation characteristics are significantly altered under the combined effects of buoyant-thermocapillary convection. Buoyancy, induced under sufficiently high Ri, was shown to predominate over the thermocapillay motion; causing the enhancement (suppression) of breakdown when the rod is warmer (cooler) than the ambient fluid. However, over small ranges of Ri, the sensitivity of the flow to surface tension gradients was clearly evidenced and results showed its full control over the occurrence and location of breakdown. In particular, detailed timewise evolution of the flow indicated that weak thermocapillary motion was sufficient to prevent the formation of toroidal patterns. These latter detach from the surface and undergo considerable size reduction while moving towards the bulk flow before vanishing. Further calculations revealed that the pattern reappears with increasing time as steady bubble type on the rod. However, in the absence of the central rod and also in the case of small rod length l, the flow evolved into steady state without any breakdown.

Keywords: buoyancy, cylinder, surface tension, toroidal vortex

Procedia PDF Downloads 334
270 Pre-Cooling Strategies for the Refueling of Hydrogen Cylinders in Vehicular Transport

Authors: C. Hall, J. Ramos, V. Ramasamy

Abstract:

Hydrocarbon-based fuel vehicles are a major contributor to air pollution due to harmful emissions produced, leading to a demand for cleaner fuel types. A leader in this pursuit is hydrogen, with its application in vehicles producing zero harmful emissions and the only by-product being water. To compete with the performance of conventional vehicles, hydrogen gas must be stored on-board of vehicles in cylinders at high pressures (35–70 MPa) and have a short refueling duration (approximately 3 mins). However, the fast-filling of hydrogen cylinders causes a significant rise in temperature due to the combination of the negative Joule-Thompson effect and the compression of the gas. This can lead to structural failure and therefore, a maximum allowable internal temperature of 85°C has been imposed by the International Standards Organization. The technological solution to tackle the issue of rapid temperature rise during the refueling process is to decrease the temperature of the gas entering the cylinder. Pre-cooling of the gas uses a heat exchanger and requires energy for its operation. Thus, it is imperative to determine the least amount of energy input that is required to lower the gas temperature for cost savings. A validated universal thermodynamic model is used to identify an energy-efficient pre-cooling strategy. The model requires negligible computational time and is applied to previously validated experimental cases to optimize pre-cooling requirements. The pre-cooling characteristics include the location within the refueling timeline and its duration. A constant pressure-ramp rate is imposed to eliminate the effects of rapid changes in mass flow rate. A pre-cooled gas temperature of -40°C is applied, which is the lowest allowable temperature. The heat exchanger is assumed to be ideal with no energy losses. The refueling of the cylinders is modeled with the pre-cooling split in ten percent time intervals. Furthermore, varying burst durations are applied in both the early and late stages of the refueling procedure. The model shows that pre-cooling in the later stages of the refuelling process is more energy-efficient than early pre-cooling. In addition, the efficiency of pre-cooling towards the end of the refueling process is independent of the pressure profile at the inlet. This leads to the hypothesis that pre-cooled gas should be applied as late as possible in the refueling timeline and at very low temperatures. The model had shown a 31% reduction in energy demand whilst achieving the same final gas temperature for a refueling scenario when pre-cooling was applied towards the end of the process. The identification of the most energy-efficient refueling approaches whilst adhering to the safety guidelines is imperative to reducing the operating cost of hydrogen refueling stations. Heat exchangers are energy-intensive and thus, reducing the energy requirement would lead to cost reduction. This investigation shows that pre-cooling should be applied as late as possible and for short durations.

Keywords: cylinder, hydrogen, pre-cooling, refueling, thermodynamic model

Procedia PDF Downloads 78
269 Breast Cancer Awareness among Female Nurses: Time to Scrub off Assumptions

Authors: Rahy Farooq, Maria Ahmad Khan, Ayesha Isani Majeed

Abstract:

Objective: The main aim of this research is to assess the knowledge, attitude and practices of female nursing staff and nursing students regarding breast cancer, to provide a baseline for monitoring trends of breast cancer awareness in them. Background: Healthcare professionals are a direct source of information for the patients and the general public as a whole. It is, therefore, essential that the information they convey be accurate and helps in building additional awareness. However, clinical experience does not influence the knowledge, attitude and practices regarding breast cancer. Nurses, being the prime part of the healthcare professionals, play a significant role and hence, their awareness regarding this pressing issue is pertinent. Lack of awareness regarding common presenting symptoms or breast cancer risk factors translates to poor breast cancer screening practices and late diagnosis. Methodology: A cross-sectional study of 280 female nurses was conducted at a tertiary care hospital in Islamabad, Pakistan. A pre-tested structured questionnaire with additional variables like cultural barriers to seeking medical help was used. The scores for outcome variables including knowledge, attitude and practices were pre-defined. Data was analyzed using SPSSv23. Results: Of the 280 participants with a mean age of 28.99±9.98 years, 142 (50.7%) were married, and 138 (49.3%) were unmarried. Mean scores were computed to be 6.14±2.93 (out of 12), 0.30±0.7 (out of 3) and 9.53±1.92 (out of 16) for knowledge, attitude and practice respectively. Using independent sample T-test, a statistically significant correlation was found when means for the score of Attitude was compared with age. With a p-value of 0.018, 117 nurses of age more than 30 years, faced more practical, financial, emotional and service barriers as compared to 163 women younger than 30 years of age. Knowledge of age-related lifetime risks was also significantly poor more in single women; with a p-value of 0.006 for identification of correct age as a risk factor and a p-value of 0.005 for correct identification of risk for development of breast cancer in the lifetime of women. By application of Chi-square test, there was a significant correlation between marital status and cultural barriers to seeking medical help, showing that single women (58.7%) shy away from talking about breast cancer considering it a taboo (p-value 0.028) whereas, more married nurses (59.2%) were apprehensive that they might be considered at fault by the society, as compared to 40.8% of single nurses. (p-value 0.038). Conclusion: Owing to the scarcity of awareness among nurses, this study recognizes the need for delivering effective information to the female nurses regarding breast cancer. Educating patients is likely to be effective if the female nurses play their part and have correct attitudes towards breast cancer practices. A better understanding of the knowledge and practices regarding breast cancer among the nursing population will enable high-risk patients to be recognized early. Therefore, we recommend arrangement of special courses and seminars for all healthcare professionals including the nursing staff.

Keywords: breast cancer, cultural barriers, kap, nurses

Procedia PDF Downloads 222
268 Assessing Acute Toxicity and Endocrine Disruption Potential of Selected Packages Internal Layers Extracts

Authors: N. Szczepanska, B. Kudlak, G. Yotova, S. Tsakovski, J. Namiesnik

Abstract:

In the scientific literature related to the widely understood issue of packaging materials designed to have contact with food (food contact materials), there is much information on raw materials used for their production, as well as their physiochemical properties, types, and parameters. However, not much attention is given to the issues concerning migration of toxic substances from packaging and its actual influence on the health of the final consumer, even though health protection and food safety are the priority tasks. The goal of this study was to estimate the impact of particular foodstuff packaging type, food production, and storage conditions on the degree of leaching of potentially toxic compounds and endocrine disruptors to foodstuffs using the acute toxicity test Microtox and XenoScreen YES YAS assay. The selected foodstuff packaging materials were metal cans used for fish storage and tetrapak. Five stimulants respectful to specific kinds of food were chosen in order to assess global migration: distilled water for aqueous foods with a pH above 4.5; acetic acid at 3% in distilled water for acidic aqueous food with pH below 4.5; ethanol at 5% for any food that may contain alcohol; dimethyl sulfoxide (DMSO) and artificial saliva were used in regard to the possibility of using it as an simulation medium. For each packaging three independent variables (temperature and contact time) factorial design simulant was performed. Xenobiotics migration from epoxy resins was studied at three different temperatures (25°C, 65°C, and 121°C) and extraction time of 12h, 48h and 2 weeks. Such experimental design leads to 9 experiments for each food simulant as conditions for each experiment are obtained by combination of temperature and contact time levels. Each experiment was run in triplicate for acute toxicity and in duplicate for estrogen disruption potential determination. Multi-factor analysis of variation (MANOVA) was used to evaluate the effects of the three main factors solvent, temperature (temperature regime for cup), contact time and their interactions on the respected dependent variable (acute toxicity or estrogen disruption potential). From all stimulants studied the most toxic were can and tetrapak lining acetic acid extracts that are indication for significant migration of toxic compounds. This migration increased with increase of contact time and temperature and justified the hypothesis that food products with low pH values cause significant damage internal resin filling. Can lining extracts of all simulation medias excluding distilled water and artificial saliva proved to contain androgen agonists even at 25°C and extraction time of 12h. For tetrapak extracts significant endocrine potential for acetic acid, DMSO and saliva were detected.

Keywords: food packaging, extraction, migration, toxicity, biotest

Procedia PDF Downloads 161
267 Buoyant Gas Dispersion in a Small Fuel Cell Enclosure: A Comparison Study Using Plain and Pressed Louvre Vent Passive Ventilation Schemes

Authors: T. Ghatauray, J. Ingram, P. Holborn

Abstract:

The transition from a ‘carbon rich’ fossil fuel dependent to a ‘sustainable’ and ‘renewable’ hydrogen based society will see the deployment of hydrogen fuel cells (HFC) in transport applications and in the generation of heat and power for buildings, as part of a decentralised power network. Many deployments will be low power HFCs for domestic combined heat and power (CHP) and commercial ‘transportable’ HFCs for environmental situations, such as lighting and telephone towers. For broad commercialisation of small fuel cells to be achieved there needs to be significant confidence in their safety in both domestic and environmental applications. Low power HFCs are housed in protective steel enclosures. Standard enclosures have plain rectangular ventilation openings intended for thermal management of electronics and not the dispersion of a buoyant gas. Degradation of the HFC or supply pipework in use could lead to a low-level leak and a build-up of hydrogen gas in the enclosure. Hydrogen’s wide flammable range (4-75%) is a significant safety concern, with ineffective enclosure ventilation having the potential to cause flammable mixtures to develop with the risk of explosion. Mechanical ventilation is effective at managing enclosure hydrogen concentrations, but drains HFC power and is vulnerable to failure. This is undesirable in low power and remote installations and reliable passive ventilation systems are preferred. Passive ventilation depends upon buoyancy driven flow, with the size, shape and position of ventilation openings critical for producing predictable flows and maintaining low buoyant gas concentrations. With environmentally sited enclosures, ventilation openings with pressed horizontal and angled louvres are preferred to protect the HFC and electronics inside. There is an economic cost to adding louvres, but also a safety concern. A question arises over whether the use of pressed louvre vents impairs enclosure passive ventilation performance, when compared to same opening area plain vents. Comparison small enclosure (0.144m³) tests of same opening area pressed louvre and plain vents were undertaken. A displacement ventilation arrangement was incorporated into the enclosure with opposing upper and lower ventilation openings. A range of vent areas were tested. Helium (used as a safe analogue for hydrogen) was released from a 4mm nozzle at the base of the enclosure to simulate a hydrogen leak at leak rates from 1 to 10 lpm. Helium sensors were used to record concentrations at eight heights in the enclosure. The enclosure was otherwise empty. These tests determined that the use of pressed and angled louvre ventilation openings on the enclosure impaired the passive ventilation flow and increased helium concentrations in the enclosure. High-level stratified buoyant gas layers were also found to be deeper than with plain vent openings and were within the flammable range. The presence of gas within the flammable range is of concern, particularly as the addition of the fuel cell and electronics in the enclosure would further reduce the available volume and increase concentrations. The opening area of louvre vents would need to be greater than equivalent plain vents to achieve comparable ventilation flows or alternative schemes would need to be considered.

Keywords: enclosure, fuel cell, helium, hydrogen safety, louvre vent, passive ventilation

Procedia PDF Downloads 254
266 Fully Autonomous Vertical Farm to Increase Crop Production

Authors: Simone Cinquemani, Lorenzo Mantovani, Aleksander Dabek

Abstract:

New technologies in agriculture are opening new challenges and new opportunities. Among these, certainly, robotics, vision, and artificial intelligence are the ones that will make a significant leap, compared to traditional agricultural techniques, possible. In particular, the indoor farming sector will be the one that will benefit the most from these solutions. Vertical farming is a new field of research where mechanical engineering can bring knowledge and know-how to transform a highly labor-based business into a fully autonomous system. The aim of the research is to develop a multi-purpose, modular, and perfectly integrated platform for crop production in indoor vertical farming. Activities will be based both on hardware development such as automatic tools to perform different activities on soil and plants, as well as research to introduce an extensive use of monitoring techniques based on machine learning algorithms. This paper presents the preliminary results of a research project of a vertical farm living lab designed to (i) develop and test vertical farming cultivation practices, (ii) introduce a very high degree of mechanization and automation that makes all processes replicable, fully measurable, standardized and automated, (iii) develop a coordinated control and management environment for autonomous multiplatform or tele-operated robots in environments with the aim of carrying out complex tasks in the presence of environmental and cultivation constraints, (iv) integrate AI-based algorithms as decision support system to improve quality production. The coordinated management of multiplatform systems still presents innumerable challenges that require a strongly multidisciplinary approach right from the design, development, and implementation phases. The methodology is based on (i) the development of models capable of describing the dynamics of the various platforms and their interactions, (ii) the integrated design of mechatronic systems able to respond to the needs of the context and to exploit the strength characteristics highlighted by the models, (iii) implementation and experimental tests performed to test the real effectiveness of the systems created, evaluate any weaknesses so as to proceed with a targeted development. To these aims, a fully automated laboratory for growing plants in vertical farming has been developed and tested. The living lab makes extensive use of sensors to determine the overall state of the structure, crops, and systems used. The possibility of having specific measurements for each element involved in the cultivation process makes it possible to evaluate the effects of each variable of interest and allows for the creation of a robust model of the system as a whole. The automation of the laboratory is completed with the use of robots to carry out all the necessary operations, from sowing to handling to harvesting. These systems work synergistically thanks to the knowledge of detailed models developed based on the information collected, which allows for deepening the knowledge of these types of crops and guarantees the possibility of tracing every action performed on each single plant. To this end, artificial intelligence algorithms have been developed to allow synergistic operation of all systems.

Keywords: automation, vertical farming, robot, artificial intelligence, vision, control

Procedia PDF Downloads 16
265 Investigating the Urban Heat Island Phenomenon in A Desert City Aiming at Sustainable Buildings

Authors: Afifa Mohammed, Gloria Pignatta, Mattheos Santamouris, Evangelia Topriska

Abstract:

Climate change is one of the global challenges that is exacerbated by the rapid growth of urbanizations. Urban Heat Island (UHI) phenomenon can be considered as an effect of the urbanization and it is responsible together with the Climate change of the overheating of urban cities and downtowns. The purpose of this paper is to quantify and perform analysis of UHI Intensity in Dubai, United Arab Emirates (UAE), through checking the relationship between the UHI and different meteorological parameters (e.g., temperature, winds speed, winds direction). Climate data were collected from three meteorological stations in Dubai (e.g., Dubai Airport - Station 1, Al-Maktoum Airport - Station 2 and Saih Al-Salem - Station 3) for a period of five years (e.g., 2014 – 2018) based upon hourly rates, and following clustering technique as one of the methodology tools of measurements. The collected data of each station were divided into six clusters upon the winds directions, either from the seaside or from the desert side, or from the coastal side which is in between both aforementioned winds sources, to investigate the relationship between temperature degrees and winds speed values through UHI measurements for Dubai Airport - Station 1 compared with the same of Al-Maktoum Airport - Station 2. In this case, the UHI value is determined by the temperature difference of both stations, where Station 1 is considered as located in an urban area and Station 2 is considered as located in a suburban area. The same UHI calculations has been applied for Al-Maktoum Airport - Station 2 and Saih Salem - Station 3 where Station 2 is considered as located in an urban area and Station 3 is considered as located in a suburban area. The performed analysis aims to investigate the relation between the two environmental parameters (e.g., Temperature and Winds Speed) and the Urban Heat Island (UHI) intensity when the wind comes from the seaside, from the desert, and the remaining directions. The analysis shows that the correlation between the temperatures with both UHI intensity (e.g., temperature difference between Dubai Airport - Station 1 and Saih Al-Salem - Station 3 and between Al-Maktoum Airport - Station 2 and Saih Al-Salem - Station 3 (through station 1 & 2) is strong and has a negative relationship when the wind is coming from the seaside comparing between the two stations 1 and 2, while the relationship is almost zero (no relation) when the wind is coming from the desert side. The relation is independent between the two parameters, e.g., temperature and UHI, on Station 2, during the same procedures, the correlation between the urban heat island UHI phenomenon and wind speed is weak for both stations when wind direction is coming from the seaside comparing the station 1 and 2, while it was found that there’s no relationship between urban heat island phenomenon and wind speed when wind direction is coming from desert side. The conclusion could be summarized saying that the wind coming from the seaside or from the desert side have a different effect on UHI, which is strongly affected by meteorological parameters. The output of this study will enable more determination of UHI phenomenon under desert climate, which will help to inform about the UHI phenomenon and intensity and extract recommendations in two main categories such as planning of new cities and designing of buildings.

Keywords: meteorological data, subtropical desert climate, urban climate, urban heat island (UHI)

Procedia PDF Downloads 120
264 Case Study Analysis of 2017 European Railway Traffic Management Incident: The Application of System for Investigation of Railway Interfaces Methodology

Authors: Sanjeev Kumar Appicharla

Abstract:

This paper presents the results of the modelling and analysis of the European Railway Traffic Management (ERTMS) safety-critical incident to raise awareness of biases in the systems engineering process on the Cambrian Railway in the UK using the RAIB 17/2019 as a primary input. The RAIB, the UK independent accident investigator, published the Report- RAIB 17/2019 giving the details of their investigation of the focal event in the form of immediate cause, causal factors, and underlying factors and recommendations to prevent a repeat of the safety-critical incident on the Cambrian Line. The Systems for Investigation of Railway Interfaces (SIRI) is the methodology used to model and analyze the safety-critical incident. The SIRI methodology uses the Swiss Cheese Model to model the incident and identify latent failure conditions (potentially less than adequate conditions) by means of the management oversight and risk tree technique. The benefits of the systems for investigation of railway interfaces methodology (SIRI) are threefold: first is that it incorporates the “Heuristics and Biases” approach advanced by 2002 Nobel laureate in Economic Sciences, Prof Daniel Kahneman, in the management oversight and risk tree technique to identify systematic errors. Civil engineering and programme management railway professionals are aware of the role “optimism bias” plays in programme cost overruns and are aware of bow tie (fault and event tree) model-based safety risk modelling techniques. However, the role of systematic errors due to “Heuristics and Biases” is not appreciated as yet. This overcomes the problems of omission of human and organizational factors from accident analysis. Second, the scope of the investigation includes all levels of the socio-technical system, including government, regulatory, railway safety bodies, duty holders, signaling firms and transport planners, and front-line staff such that lessons are learned at the decision making and implementation level as well. Third, the author’s past accident case studies are supplemented with research pieces of evidence drawn from the practitioner's and academic researchers’ publications as well. This is to discuss the role of system thinking to improve the decision-making and risk management processes and practices in the IEC 15288 systems engineering standard and in the industrial context such as the GB railways and artificial intelligence (AI) contexts as well.

Keywords: accident analysis, AI algorithm internal audit, bounded rationality, Byzantine failures, heuristics and biases approach

Procedia PDF Downloads 177
263 Inconsistent Effects of Landscape Heterogeneity on Animal Diversity in an Agricultural Mosaic: A Multi-Scale and Multi-Taxon Investigation

Authors: Chevonne Reynolds, Robert J. Fletcher, Jr, Celine M. Carneiro, Nicole Jennings, Alison Ke, Michael C. LaScaleia, Mbhekeni B. Lukhele, Mnqobi L. Mamba, Muzi D. Sibiya, James D. Austin, Cebisile N. Magagula, Themba’alilahlwa Mahlaba, Ara Monadjem, Samantha M. Wisely, Robert A. McCleery

Abstract:

A key challenge for the developing world is reconciling biodiversity conservation with the growing demand for food. In these regions, agriculture is typically interspersed among other land-uses creating heterogeneous landscapes. A primary hypothesis for promoting biodiversity in agricultural landscapes is the habitat heterogeneity hypothesis. While there is evidence that landscape heterogeneity positively influences biodiversity, the application of this hypothesis is hindered by a need to determine which components of landscape heterogeneity drive these effects and at what spatial scale(s). Additionally, whether diverse taxonomic groups are similarly affected is central for determining the applicability of this hypothesis as a general conservation strategy in agricultural mosaics. Two major components of landscape heterogeneity are compositional and configurational heterogeneity. Disentangling the roles of each component is important for biodiversity conservation because each represents different mechanisms underpinning variation in biodiversity. We identified a priori independent gradients of compositional and configurational landscape heterogeneity within an extensive agricultural mosaic in north-eastern Swaziland. We then tested how bird, dung beetle, ant and meso-carnivore diversity responded to compositional and configurational heterogeneity across six different spatial scales. To determine if a general trend could be observed across multiple taxa, we also tested which component and spatial scale was most influential across all taxonomic groups combined, Compositional, not configurational, heterogeneity explained diversity in each taxonomic group, with the exception of meso-carnivores. Bird and ant diversity was positively correlated with compositional heterogeneity at fine spatial scales < 1000 m, whilst dung beetle diversity was negatively correlated to compositional heterogeneity at broader spatial scales > 1500 m. Importantly, because of these contrasting effects across taxa, there was no effect of either component of heterogeneity on the combined taxonomic diversity at any spatial scale. The contrasting responses across taxonomic groups exemplify the difficulty in implementing effective conservation strategies that meet the requirements of diverse taxa. To promote diverse communities across a range of taxa, conservation strategies must be multi-scaled and may involve different strategies at varying scales to offset the contrasting influences of compositional heterogeneity. A diversity of strategies are likely key to conserving biodiversity in agricultural mosaics, and we have demonstrated that a landscape management strategy that only manages for heterogeneity at one particular scale will likely fall short of management objectives.

Keywords: agriculture, biodiversity, composition, configuration, heterogeneity

Procedia PDF Downloads 234
262 Development of Adaptive Proportional-Integral-Derivative Feeding Mechanism for Robotic Additive Manufacturing System

Authors: Andy Alubaidy

Abstract:

In this work, a robotic additive manufacturing system (RAMS) that is capable of three-dimensional (3D) printing in six degrees of freedom (DOF) with very high accuracy and virtually on any surface has been designed and built. One of the major shortcomings in existing 3D printer technology is the limitation to three DOF, which results in prolonged fabrication time. Depending on the techniques used, it usually takes at least two hours to print small objects and several hours for larger objects. Another drawback is the size of the printed objects, which is constrained by the physical dimensions of most low-cost 3D printers, which are typically small. In such cases, large objects are produced by dividing them into smaller components that fit the printer’s workable area. They are then glued, bonded or otherwise attached to create the required object. Another shortcoming is material constraints and the need to fabricate a single part using different materials. With the flexibility of a six-DOF robot, the RAMS has been designed to overcome these problems. A feeding mechanism using an adaptive Proportional-Integral-Derivative (PID) controller is utilized along with a national instrument compactRIO (NI cRIO), an ABB robot, and off-the-shelf sensors. The RAMS have the ability to 3D print virtually anywhere in six degrees of freedom with very high accuracy. It is equipped with an ABB IRB 120 robot to achieve this level of accuracy. In order to convert computer-aided design (CAD) files to digital format that is acceptable to the robot, Hypertherm Robotic Software Inc.’s state-of-the-art slicing software called “ADDMAN” is used. ADDMAN is capable of converting any CAD file into RAPID code (the programing language for ABB robots). The robot uses the generated code to perform the 3D printing. To control the entire process, National Instrument (NI) compactRIO (cRio 9074), is connected and communicated with the robot and a feeding mechanism that is designed and fabricated. The feeding mechanism consists of two major parts, cold-end and hot-end. The cold-end consists of what is conventionally known as an extruder. Typically, a stepper-motor is used to control the push on the material, however, for optimum control, a DC motor is used instead. The hot-end consists of a melt-zone, nozzle, and heat-brake. The melt zone ensures a thorough melting effect and consistent output from the nozzle. Nozzles are made of brass for thermo-conductivity while the melt-zone is comprised of a heating block and a ceramic heating cartridge to transfer heat to the block. The heat-brake ensures that there is no heat creep-up effect as this would swell the material and prevent consistent extrusion. A control system embedded in the cRio is developed using NI Labview which utilizes adaptive PID to govern the heating cartridge in conjunction with a thermistor. The thermistor sends temperature feedback to the cRio, which will issue heat increase or decrease based on the system output. Since different materials have different melting points, our system will allow us to adjust the temperature and vary the material.

Keywords: robotic, additive manufacturing, PID controller, cRIO, 3D printing

Procedia PDF Downloads 196
261 Comparison of Two Home Sleep Monitors Designed for Self-Use

Authors: Emily Wood, James K. Westphal, Itamar Lerner

Abstract:

Background: Polysomnography (PSG) recordings are regularly used in research and clinical settings to study sleep and sleep-related disorders. Typical PSG studies are conducted in professional laboratories and performed by qualified researchers. However, the number of sleep labs worldwide is disproportionate to the increasing number of individuals with sleep disorders like sleep apnea and insomnia. Consequently, there is a growing need to supply cheaper yet reliable means to measure sleep, preferably autonomously by subjects in their own home. Over the last decade, a variety of devices for self-monitoring of sleep became available in the market; however, very few have been directly validated against PSG to demonstrate their ability to perform reliable automatic sleep scoring. Two popular mobile EEG-based systems that have published validation results, the DREEM 3 headband and the Z-Machine, have never been directly compared one to the other by independent researchers. The current study aimed to compare the performance of DREEM 3 and the Z-Machine to help investigators and clinicians decide which of these devices may be more suitable for their studies. Methods: 26 participants have completed the study for credit or monetary compensation. Exclusion criteria included any history of sleep, neurological or psychiatric disorders. Eligible participants arrived at the lab in the afternoon and received the two devices. They then spent two consecutive nights monitoring their sleep at home. Participants were also asked to keep a sleep log, indicating the time they fell asleep, woke up, and the number of awakenings occurring during the night. Data from both devices, including detailed sleep hypnograms in 30-second epochs (differentiating Wake, combined N1/N2, N3; and Rapid Eye Movement sleep), were extracted and aligned upon retrieval. For analysis, the number of awakenings each night was defined as four or more consecutive wake epochs between sleep onset and termination. Total sleep time (TST) and the number of awakenings were compared to subjects’ sleep logs to measure consistency with the subjective reports. In addition, the sleep scores from each device were compared epoch-by-epoch to calculate the agreement between the two devices using Cohen’s Kappa. All analysis was performed using Matlab 2021b and SPSS 27. Results/Conclusion: Subjects consistently reported longer times spent asleep than the time reported by each device (M= 448 minutes for sleep logs compared to M= 406 and M= 345 minutes for the DREEM and Z-Machine, respectively; both ps<0.05). Linear correlations between the sleep log and each device were higher for the DREEM than the Z-Machine for both TST and the number of awakenings, and, likewise, the mean absolute bias between the sleep logs and each device was higher for the Z-Machine for both TST (p<0.001) and awakenings (p<0.04). There was some indication that these effects were stronger for the second night compared to the first night. Epoch-by-epoch comparisons showed that the main discrepancies between the devices were for detecting N2 and REM sleep, while N3 had a high agreement. Overall, the DREEM headband seems superior for reliably scoring sleep at home.

Keywords: DREEM, EEG, seep monitoring, Z-machine

Procedia PDF Downloads 85
260 Assessing Socio-economic Impacts of Arsenic and Iron Contamination in Groundwater: Feasibility of Rainwater Harvesting in Amdanga Block, North 24 Parganas, West Bengal, India

Authors: Rajkumar Ghosh

Abstract:

The present study focuses on conducting a socio-economic assessment of groundwater contamination by arsenic and iron and explores the feasibility of rainwater harvesting (RWH) as an alternative water source in the Amdanga Block of North 24 Parganas, West Bengal, India. The region is plagued by severe groundwater contamination, primarily due to excessive concentrations of arsenic and iron, which pose significant health risks to the local population. The study utilizes a mixed-methods approach, combining quantitative analysis of water samples collected from different locations within the Amdanga Block and socio-economic surveys conducted among the affected communities. The results reveal alarmingly high levels of arsenic and iron contamination in the groundwater, surpassing the World Health Organization (WHO) and Indian government's permissible limits. This contamination significantly impacts the health and well-being of the local population, leading to a range of health issues such as skin The water samples are analyzed for arsenic and iron levels, while the surveys gather data on water usage patterns, health conditions, and socio-economic factors. lesions, respiratory disorders, and gastrointestinal problems. Furthermore, the socio-economic assessment highlights the vulnerability of the affected communities due to limited access to safe drinking water. The findings reveal the adverse socio-economic implications, including increased medical expenditures, reduced productivity, and compromised educational opportunities. To address these challenges, the study explores the feasibility of rainwater harvesting as an alternative source of clean water. RWH systems have the potential to mitigate groundwater contamination by providing a sustainable and independent water supply. The assessment includes evaluating the rainwater availability, analyzing the infrastructure requirements, and estimating the potential benefits and challenges associated with RWH implementation in the study area. The findings of this study contribute to a comprehensive understanding of the socio-economic impact of groundwater contamination by arsenic and iron, emphasizing the urgency to address this critical issue in the Amdanga Block. The feasibility assessment of rainwater harvesting serves as a practical solution to ensure a safe and sustainable water supply, reducing the dependency on contaminated groundwater sources. The study's results can inform policymakers, researchers, and local stakeholders in implementing effective mitigation measures and promoting the adoption of rainwater harvesting as a viable alternative in similar arsenic and iron-contaminated regions.

Keywords: contamination, rainwater harvesting, groundwater, sustainable water supply

Procedia PDF Downloads 78
259 Collaborative Management Approach for Logistics Flow Management of Cuban Medicine Supply Chain

Authors: Ana Julia Acevedo Urquiaga, Jose A. Acevedo Suarez, Ana Julia Urquiaga Rodriguez, Neyfe Sablon Cossio

Abstract:

Despite the progress made in logistics and supply chains fields, it is unavoidable the development of business models that use efficiently information to facilitate the integrated logistics flows management between partners. Collaborative management is an important tool for materializing the cooperation between companies, as a way to achieve the supply chain efficiency and effectiveness. The first face of this research was a comprehensive analysis of the collaborative planning on the Cuban companies. It is evident that they have difficulties in supply chains planning where production, supplies and replenishment planning are independent tasks, as well as logistics and distribution operations. Large inventories generate serious financial and organizational problems for entities, demanding increasing levels of working capital that cannot be financed. Problems were found in the efficient application of Information and Communication Technology on business management. The general objective of this work is to develop a methodology that allows the deployment of a planning and control system in a coordinated way on the medicine’s logistics system in Cuba. To achieve these objectives, several mechanisms of supply chain coordination, mathematical programming models, and other management techniques were analyzed to meet the requirements of collaborative logistics management in Cuba. One of the findings is the practical and theoretical inadequacies of the studied models to solve the current situation of the Cuban logistics systems management. To contribute to the tactical-operative management of logistics, the Collaborative Logistics Flow Management Model (CLFMM) is proposed as a tool for the balance of cycles, capacities, and inventories, always to meet the final customers’ demands in correspondence with the service level expected by these. The CLFMM has as center the supply chain planning and control system as a unique information system, which acts on the processes network. The development of the model is based on the empirical methods of analysis-synthesis and the study cases. Other finding is the demonstration of the use of a single information system to support the supply chain logistics management, allows determining the deadlines and quantities required in each process. This ensures that medications are always available to patients and there are no faults that put the population's health at risk. The simulation of planning and control with the CLFMM in medicines such as dipyrone and chlordiazepoxide, during 5 months of 2017, permitted to take measures to adjust the logistic flow, eliminate delayed processes and avoid shortages of the medicines studied. As a result, the logistics cycle efficiency can be increased to 91%, the inventory rotation would increase, and this results in a release of financial resources.

Keywords: collaborative management, medicine logistic system, supply chain planning, tactical-operative planning

Procedia PDF Downloads 155
258 Benefits of Using Social Media and Collaborative Online Platforms in PBL

Authors: Susanna Graziano, Lydia Krstic Ward

Abstract:

The purpose of this presentation is to demonstrate the steps of using multimedia and collaborative platforms in project-based learning. The presentation will demonstrate the stages of the learning project with various components of independent and collaborative learning, where students research the topic, share information, prepare a survey, use social media (Facebook, Instagram, WhasApp) and collaborative platforms (wikispaces.com and Google docs) to collect, analyze and process data, then produce reports and logos to be displayed as a final product. At the beginning of the presentation participants will answer a questionnaire about project based learning and share their experience on using social media, real–world project work and collaborative learning. Using a PPP, the presentation will walk participants through the steps of a completed project where tertiary education students are involved in putting together a multimedia campaign for safe driving in Kuwait. The research component of the project entails taking a holistic view on the problem of the high death rate in traffic accidents. The final goal of the project is to lead students to raise public awareness about the importance of safe driving. The project steps involve using the social media and collaborative platforms for collecting data and sharing the required materials to be used in the final product – a display of written reports, slogans and videos, as well as oral presentations. The same structure can be used to organize a multimedia campaign focusing on other issues, whilst scaffolding on students’ ability to brainstorm, retrieve information, organize it and engage in collaborative/ cooperative learning whilst being immersed in content-based learning as well as in authentic tasks. More specifically, the project we carried out at Box Hill College was a real-world one and involved a multimedia Campaign for Safe Driving since reckless driving is one of the major problems in the country. The idea for the whole project started by a presentation given by a board member of the Kuwaiti Society for Traffic Safety who was invited to college and spoke about: • Driving laws in the country, • What causes car accidents, • Driving safety tips. The principal goal of this project was to let students consider problems of traffic in Kuwait from different points of view. They also had to address the number and causes of accidents, evaluate the effectiveness of the local traffic law in order to send a warning about the importance of safe driving and, finally, suggest ways of its improvement. Benefits included: • Engagement, • Autonomy, • Motivation, • Content knowledge, • Language mastery, • Enhanced critical thinking, • Increased metacognitive awareness, • Improved social skills, • Authentic experience.

Keywords: social media, online learning platforms, collaborative platforms, project based learning

Procedia PDF Downloads 405
257 ChatGPT Performs at the Level of a Third-Year Orthopaedic Surgery Resident on the Orthopaedic In-training Examination

Authors: Diane Ghanem, Oscar Covarrubias, Michael Raad, Dawn LaPorte, Babar Shafiq

Abstract:

Introduction: Standardized exams have long been considered a cornerstone in measuring cognitive competency and academic achievement. Their fixed nature and predetermined scoring methods offer a consistent yardstick for gauging intellectual acumen across diverse demographics. Consequently, the performance of artificial intelligence (AI) in this context presents a rich, yet unexplored terrain for quantifying AI's understanding of complex cognitive tasks and simulating human-like problem-solving skills. Publicly available AI language models such as ChatGPT have demonstrated utility in text generation and even problem-solving when provided with clear instructions. Amidst this transformative shift, the aim of this study is to assess ChatGPT’s performance on the orthopaedic surgery in-training examination (OITE). Methods: All 213 OITE 2021 web-based questions were retrieved from the AAOS-ResStudy website. Two independent reviewers copied and pasted the questions and response options into ChatGPT Plus (version 4.0) and recorded the generated answers. All media-containing questions were flagged and carefully examined. Twelve OITE media-containing questions that relied purely on images (clinical pictures, radiographs, MRIs, CT scans) and could not be rationalized from the clinical presentation were excluded. Cohen’s Kappa coefficient was used to examine the agreement of ChatGPT-generated responses between reviewers. Descriptive statistics were used to summarize the performance (% correct) of ChatGPT Plus. The 2021 norm table was used to compare ChatGPT Plus’ performance on the OITE to national orthopaedic surgery residents in that same year. Results: A total of 201 were evaluated by ChatGPT Plus. Excellent agreement was observed between raters for the 201 ChatGPT-generated responses, with a Cohen’s Kappa coefficient of 0.947. 45.8% (92/201) were media-containing questions. ChatGPT had an average overall score of 61.2% (123/201). Its score was 64.2% (70/109) on non-media questions. When compared to the performance of all national orthopaedic surgery residents in 2021, ChatGPT Plus performed at the level of an average PGY3. Discussion: ChatGPT Plus is able to pass the OITE with a satisfactory overall score of 61.2%, ranking at the level of third-year orthopaedic surgery residents. More importantly, it provided logical reasoning and justifications that may help residents grasp evidence-based information and improve their understanding of OITE cases and general orthopaedic principles. With further improvements, AI language models, such as ChatGPT, may become valuable interactive learning tools in resident education, although further studies are still needed to examine their efficacy and impact on long-term learning and OITE/ABOS performance.

Keywords: artificial intelligence, ChatGPT, orthopaedic in-training examination, OITE, orthopedic surgery, standardized testing

Procedia PDF Downloads 63
256 Reliable and Error-Free Transmission through Multimode Polymer Optical Fibers in House Networks

Authors: Tariq Ahamad, Mohammed S. Al-Kahtani, Taisir Eldos

Abstract:

Optical communications technology has made enormous and steady progress for several decades, providing the key resource in our increasingly information-driven society and economy. Much of this progress has been in finding innovative ways to increase the data carrying capacity of a single optical fiber. In this research article we have explored basic issues in terms of security and reliability for secure and reliable information transfer through the fiber infrastructure. Conspicuously, one potentially enormous source of improvement has however been left untapped in these systems: fibers can easily support hundreds of spatial modes, but today’s commercial systems (single-mode or multi-mode) make no attempt to use these as parallel channels for independent signals. Bandwidth, performance, reliability, cost efficiency, resiliency, redundancy, and security are some of the demands placed on telecommunications today. Since its initial development, fiber optic systems have had the advantage of most of these requirements over copper-based and wireless telecommunications solutions. The largest obstacle preventing most businesses from implementing fiber optic systems was cost. With the recent advancements in fiber optic technology and the ever-growing demand for more bandwidth, the cost of installing and maintaining fiber optic systems has been reduced dramatically. With so many advantages, including cost efficiency, there will continue to be an increase of fiber optic systems replacing copper-based communications. This will also lead to an increase in the expertise and the technology needed to tap into fiber optic networks by intruders. As ever before, all technologies have been subject to hacking and criminal manipulation, fiber optics is no exception. Researching fiber optic security vulnerabilities suggests that not everyone who is responsible for their networks security is aware of the different methods that intruders use to hack virtually undetected into fiber optic cables. With millions of miles of fiber optic cables stretching across the globe and carrying information including but certainly not limited to government, military, and personal information, such as, medical records, banking information, driving records, and credit card information; being aware of fiber optic security vulnerabilities is essential and critical. Many articles and research still suggest that fiber optics is expensive, impractical and hard to tap. Others argue that it is not only easily done, but also inexpensive. This paper will briefly discuss the history of fiber optics, explain the basics of fiber optic technologies and then discuss the vulnerabilities in fiber optic systems and how they can be better protected. Knowing the security risks and knowing the options available may save a company a lot embarrassment, time, and most importantly money.

Keywords: in-house networks, fiber optics, security risk, money

Procedia PDF Downloads 402
255 Spatial Climate Changes in the Province of Macerata, Central Italy, Analyzed by GIS Software

Authors: Matteo Gentilucci, Marco Materazzi, Gilberto Pambianchi

Abstract:

Climate change is an increasingly central issue in the world, because it affects many of human activities. In this context regional studies are of great importance because they sometimes differ from the general trend. This research focuses on a small area of central Italy which overlooks the Adriatic Sea, the province of Macerata. The aim is to analyze space-based climate changes, for precipitation and temperatures, in the last 3 climatological standard normals (1961-1990; 1971-2000; 1981-2010) through GIS software. The data collected from 30 weather stations for temperature and 61 rain gauges for precipitation were subject to quality controls: validation and homogenization. These data were fundamental for the spatialization of the variables (temperature and precipitation) through geostatistical techniques. To assess the best geostatistical technique for interpolation, the results of cross correlation were used. The co-kriging method with altitude as independent variable produced the best cross validation results for all time periods, among the methods analysed, with 'root mean square error standardized' close to 1, 'mean standardized error' close to 0, 'average standard error' and 'root mean square error' with similar values. The maps resulting from the analysis were compared by subtraction between rasters, producing 3 maps of annual variation and three other maps for each month of the year (1961/1990-1971/2000; 1971/2000-1981/2010; 1961/1990-1981/2010). The results show an increase in average annual temperature of about 0.1°C between 1961-1990 and 1971-2000 and 0.6 °C between 1961-1990 and 1981-2010. Instead annual precipitation shows an opposite trend, with an average difference from 1961-1990 to 1971-2000 of about 35 mm and from 1961-1990 to 1981-2010 of about 60 mm. Furthermore, the differences in the areas have been highlighted with area graphs and summarized in several tables as descriptive analysis. In fact for temperature between 1961-1990 and 1971-2000 the most areally represented frequency is 0.08°C (77.04 Km² on a total of about 2800 km²) with a kurtosis of 3.95 and a skewness of 2.19. Instead, the differences for temperatures from 1961-1990 to 1981-2010 show a most areally represented frequency of 0.83 °C, with -0.45 as kurtosis and 0.92 as skewness (36.9 km²). Therefore it can be said that distribution is more pointed for 1961/1990-1971/2000 and smoother but more intense in the growth for 1961/1990-1981/2010. In contrast, precipitation shows a very similar shape of distribution, although with different intensities, for both variations periods (first period 1961/1990-1971/2000 and second one 1961/1990-1981/2010) with similar values of kurtosis (1st = 1.93; 2nd = 1.34), skewness (1st = 1.81; 2nd = 1.62 for the second) and area of the most represented frequency (1st = 60.72 km²; 2nd = 52.80 km²). In conclusion, this methodology of analysis allows the assessment of small scale climate change for each month of the year and could be further investigated in relation to regional atmospheric dynamics.

Keywords: climate change, GIS, interpolation, co-kriging

Procedia PDF Downloads 108
254 Multi-Label Approach to Facilitate Test Automation Based on Historical Data

Authors: Warda Khan, Remo Lachmann, Adarsh S. Garakahally

Abstract:

The increasing complexity of software and its applicability in a wide range of industries, e.g., automotive, call for enhanced quality assurance techniques. Test automation is one option to tackle the prevailing challenges by supporting test engineers with fast, parallel, and repetitive test executions. A high degree of test automation allows for a shift from mundane (manual) testing tasks to a more analytical assessment of the software under test. However, a high initial investment of test resources is required to establish test automation, which is, in most cases, a limitation to the time constraints provided for quality assurance of complex software systems. Hence, a computer-aided creation of automated test cases is crucial to increase the benefit of test automation. This paper proposes the application of machine learning for the generation of automated test cases. It is based on supervised learning to analyze test specifications and existing test implementations. The analysis facilitates the identification of patterns between test steps and their implementation with test automation components. For the test case generation, this approach exploits historical data of test automation projects. The identified patterns are the foundation to predict the implementation of unknown test case specifications. Based on this support, a test engineer solely has to review and parameterize the test automation components instead of writing them manually, resulting in a significant time reduction for establishing test automation. Compared to other generation approaches, this ML-based solution can handle different writing styles, authors, application domains, and even languages. Furthermore, test automation tools require expert knowledge by means of programming skills, whereas this approach only requires historical data to generate test cases. The proposed solution is evaluated using various multi-label evaluation criteria (EC) and two small-sized real-world systems. The most prominent EC is ‘Subset Accuracy’. The promising results show an accuracy of at least 86% for test cases, where a 1:1 relationship (Multi-Class) between test step specification and test automation component exists. For complex multi-label problems, i.e., one test step can be implemented by several components, the prediction accuracy is still at 60%. It is better than the current state-of-the-art results. It is expected the prediction quality to increase for larger systems with respective historical data. Consequently, this technique facilitates the time reduction for establishing test automation and is thereby independent of the application domain and project. As a work in progress, the next steps are to investigate incremental and active learning as additions to increase the usability of this approach, e.g., in case labelled historical data is scarce.

Keywords: machine learning, multi-class, multi-label, supervised learning, test automation

Procedia PDF Downloads 104
253 A Comparison of Methods for Estimating Dichotomous Treatment Effects: A Simulation Study

Authors: Jacqueline Y. Thompson, Sam Watson, Lee Middleton, Karla Hemming

Abstract:

Introduction: The odds ratio (estimated via logistic regression) is a well-established and common approach for estimating covariate-adjusted binary treatment effects when comparing a treatment and control group with dichotomous outcomes. Its popularity is primarily because of its stability and robustness to model misspecification. However, the situation is different for the relative risk and risk difference, which are arguably easier to interpret and better suited to specific designs such as non-inferiority studies. So far, there is no equivalent, widely acceptable approach to estimate an adjusted relative risk and risk difference when conducting clinical trials. This is partly due to the lack of a comprehensive evaluation of available candidate methods. Methods/Approach: A simulation study is designed to evaluate the performance of relevant candidate methods to estimate relative risks to represent conditional and marginal estimation approaches. We consider the log-binomial, generalised linear models (GLM) with iteratively weighted least-squares (IWLS) and model-based standard errors (SE); log-binomial GLM with convex optimisation and model-based SEs; log-binomial GLM with convex optimisation and permutation tests; modified-Poisson GLM IWLS and robust SEs; log-binomial generalised estimation equations (GEE) and robust SEs; marginal standardisation and delta method SEs; and marginal standardisation and permutation test SEs. Independent and identically distributed datasets are simulated from a randomised controlled trial to evaluate these candidate methods. Simulations are replicated 10000 times for each scenario across all possible combinations of sample sizes (200, 1000, and 5000), outcomes (10%, 50%, and 80%), and covariates (ranging from -0.05 to 0.7) representing weak, moderate or strong relationships. Treatment effects (ranging from 0, -0.5, 1; on the log-scale) will consider null (H0) and alternative (H1) hypotheses to evaluate coverage and power in realistic scenarios. Performance measures (bias, mean square error (MSE), relative efficiency, and convergence rates) are evaluated across scenarios covering a range of sample sizes, event rates, covariate prognostic strength, and model misspecifications. Potential Results, Relevance & Impact: There are several methods for estimating unadjusted and adjusted relative risks. However, it is unclear which method(s) is the most efficient, preserves type-I error rate, is robust to model misspecification, or is the most powerful when adjusting for non-prognostic and prognostic covariates. GEE estimations may be biased when the outcome distributions are not from marginal binary data. Also, it seems that marginal standardisation and convex optimisation may perform better than GLM IWLS log-binomial.

Keywords: binary outcomes, statistical methods, clinical trials, simulation study

Procedia PDF Downloads 91
252 Survey of the Literacy by Radio Project as an Innovation in Literacy Promotion in Nigeria

Authors: Stella Chioma Nwizu

Abstract:

The National Commission for Adult and Non Formal Education (NMEC) in Nigeria is charged with the reduction of illiteracy rate through the development, monitoring, and supervision of literacy programmes in Nigeria. In spite of various efforts by NMEC to reduce illiteracy, literature still shows that the illiteracy rate is still high. According to NMEC/UNICEF, about 60 million Nigerians are non-literate, and nearly two thirds of them are women. This situation forced the government to search for innovative and better approaches to literacy promotion and delivery. The literacy by radio project was adopted as an innovative intervention to literacy delivery in Nigeria because the radio is the cheapest and most easily affordable medium for non-literates. The project aimed at widening access to literacy programmes for the non-literate marginalized and disadvantaged groups in Nigeria by taking literacy programmes to their door steps. The literacy by radio has worked perfectly well in non-literacy reduction in Cuba. This innovative intervention of literacy by radio is anchored on the diffusion of innovation theory by Rogers. The literacy by radio has been going on for fifteen years and the efficacy and contributions of this innovation need to be investigated. Thus, the purpose of this research is to review the contributions of the literacy by radio in Nigeria. The researcher adopted the survey research design for the study. The population for the study consisted of 2,706 participants and 47 facilitators of the literacy by radio programme in the 10 pilot states in Nigeria. A sample of four states made up of 302 participants and eight facilitators were used for the study. Information was collected through Focus Group Discussion (FGD), interviews and content analysis of official documents. The data were analysed qualitatively to review the contributions of literacy by radio project and determine the efficacy of this innovative approach in facilitating literacy in Nigeria. Results from the field experience showed, among others, that more non-literates have better access to literacy programmes through this innovative approach. The pilot project was 88% successful; not less than 2,110 adults were made literate through the literacy by radio project in 2017. However, lack of enthusiasm and commitment on the part of the technical committee and facilitators due to non-payment of honorarium, poor signals from radio stations, interruption of lectures with adverts, low community involvement in decision making in the project are challenges to the success rate of the project. The researcher acknowledges the need to customize all materials and broadcasts in all the dialects of the participants and the inclusion of more civil rights, environmental protection and agricultural skills into the project. The study recommends among others, improved and timely funding of the project by the Federal Government to enable NMEC to fulfill her obligations towards the greater success of the programme, setting up of independent radio stations for airing the programmes and proper monitoring and evaluation of the project by NMEC and State Agencies for greater effectiveness. In an era of the knowledge-driven economy, no one should be allowed to get saddled with the weight of illiteracy.

Keywords: innovative approach, literacy, project, radio, survey

Procedia PDF Downloads 43
251 The Psychological and Subjective Well-being of Ethiopian adults: Correlates, Explanations, and Cross-Cultural Constructions

Authors: Kassahun Tilahun

Abstract:

The purpose of the study was two-fold: to examine the socio-demographic and psychological predictors of well-being and formulate a socio-culturally sound approach explaining the meaning and experience of psychological well-being among Ethiopian adults. Ryan and Deci’s Self-Determination Theory was duly considered as a theoretical framework of the study. The study followed a sequential explanatory mixed method design. Both quantitative and qualitative data were obtained, via scales and open-ended questionnaires, from 438 civil servants working in Addis Ababa. 30 interviews were also conducted to gain further information. An in-depth analysis of the reliability and validity of instruments was made before employing them to the main study. The results showed that adults were better off in both their scores of psychological and subjective well-being. Besides, adults’ well-being was found to be quite a function of their gender, age, marital status, educational level and household income. Males had a healthier psychological well-being status than females, where as females were better in their subjective well-being. A significant difference in psychological well-being was also observed between emerging and young adults, in favor of the young; and between cohabitated and married adults, married being advantageous. A significant difference in subjective well-being measures was also noticed among single, cohabitated and married adults, in favor of the married adults in all measures. The finding revealed that happiness level of adults decrease as their educational status increases while the reverse is true to psychological well-being. Besides, as adults’ household income boosts, so do their psychological well-being and satisfaction in life. The regression analysis also produced significant independent contributions of household income to overall well-being of adults. As such, subjective well-being was significantly predicted by dummy variable of sex and marital status. Likewise, the agreeableness, conscientiousness, neuroticism and openness dimensions of personality were notable significant predictors of adults’ psychological well-being where as extraversion and agreeableness were significant predictors of their subjective well-being. Religiosity was also a significant predictor of adults’ psychological well-being. Besides, adults’ well-being was significantly predicted by the interaction between conscientiousness and religiosity. From goal pursuit dimensions, attainment of extrinsic life goals was a significant predictor of both psychological and subjective well-being. Importance and attainment of intrinsic life goals also significantly predicts adults’ psychological well-being. Finally, the subjective well-being of adults was significantly predicted by environmental mastery, positive relations with others, self-acceptance and overall psychological well-being scores of adults. The thematic analysis identified five major categories of themes, which are essential in explaining the psychological well-being of Ethiopian adults. These were; socio-cultural harmony, social cohesion, security, competence and accomplishment, and the self. Detailed discussion on the rational for including these themes was made and appropriate implications were proposed. Researchers are encouraged to expand the findings of this research and in turn develop a suitable approach taping the psychological well-being of adults living in countries like Ethiopia.

Keywords: psychological well-being, subjective well-being, adulthood, Ethiopia

Procedia PDF Downloads 514