Search results for: comprehensive feature extraction
4937 Research on the Influence of Robot Teaching on the Creativity of Primary and Secondary School Students under the Background of STEM Education
Authors: Chu Liu
Abstract:
With the development of society and the changes of the times, the requirements for the cultivation of learners are different. In the 21st century, STEM education has become a boom in the development of education in various countries, aiming to improve the comprehensive ability of learners in science, technology, engineering, and mathematics. The rise of robot education provides an effective way for STEM education to cultivate computational thinking ability, interdisciplinary ability, problem-solving ability, and teamwork ability. Although robot education has been developed in China for several years, it still lacks a standard curriculum system. This article uses programming software as a platform, through the research and analysis of 'Basic Education Information Technology Curriculum Standards (2012 Edition)', combines with the actual learning situation of learners, tries to conduct teaching project design research, and aims at providing references for the teaching ideas and method of robot education courses. In contemporary society, technological advances increasingly require creativity. Innovative comprehensive talents urgently need a radical and effective education reform to keep up with social changes. So in this context, robot teaching design can be used for students. The tendency of creativity to influence is worth to be verified.Keywords: STEM education, robot teaching, primary and secondary school students, tendency of creativity
Procedia PDF Downloads 1204936 A Comprehensive Approach in Calculating the Impact of the Ground on Radiated Electromagnetic Fields Due to Lightning
Authors: Lahcene Boukelkoul
Abstract:
The influence of finite ground conductivity is of great importance in calculating the induced voltages from the radiated electromagnetic fields due to lightning. In this paper, we try to give a comprehensive approach to calculate the impact of the ground on the radiated electromagnetic fields to lightning. The vertical component of lightning electric field is calculated with a reasonable approximation assuming a perfectly conducting ground in case the observation point does not exceed a few kilometres from the lightning channel. However, for distant observation points the radiated vertical component of lightning electric field is attenuated due finitely conducting ground. The attenuation is calculated using the expression elaborated for both low and high frequencies. The horizontal component of the electric field, however, is more affected by a finite conductivity of a ground. Besides, the contribution of the horizontal component of the electric field, to induced voltages on an overhead transmission line, is greater than that of the vertical component. Therefore, the calculation of the horizontal electric field is great concern for the simulation of lightning-induced voltages. For field to transmission lines coupling the ground impedance is calculated for early time behaviour and for low frequency range.Keywords: power engineering, radiated electromagnetic fields, lightning-induced voltages, lightning electric field
Procedia PDF Downloads 4044935 A New Method Separating Relevant Features from Irrelevant Ones Using Fuzzy and OWA Operator Techniques
Authors: Imed Feki, Faouzi Msahli
Abstract:
Selection of relevant parameters from a high dimensional process operation setting space is a problem frequently encountered in industrial process modelling. This paper presents a method for selecting the most relevant fabric physical parameters for each sensory quality feature. The proposed relevancy criterion has been developed using two approaches. The first utilizes a fuzzy sensitivity criterion by exploiting from experimental data the relationship between physical parameters and all the sensory quality features for each evaluator. Next an OWA aggregation procedure is applied to aggregate the ranking lists provided by different evaluators. In the second approach, another panel of experts provides their ranking lists of physical features according to their professional knowledge. Also by applying OWA and a fuzzy aggregation model, the data sensitivity-based ranking list and the knowledge-based ranking list are combined using our proposed percolation technique, to determine the final ranking list. The key issue of the proposed percolation technique is to filter automatically and objectively the relevant features by creating a gap between scores of relevant and irrelevant parameters. It permits to automatically generate threshold that can effectively reduce human subjectivity and arbitrariness when manually choosing thresholds. For a specific sensory descriptor, the threshold is defined systematically by iteratively aggregating (n times) the ranking lists generated by OWA and fuzzy models, according to a specific algorithm. Having applied the percolation technique on a real example, of a well known finished textile product especially the stonewashed denims, usually considered as the most important quality criteria in jeans’ evaluation, we separate the relevant physical features from irrelevant ones for each sensory descriptor. The originality and performance of the proposed relevant feature selection method can be shown by the variability in the number of physical features in the set of selected relevant parameters. Instead of selecting identical numbers of features with a predefined threshold, the proposed method can be adapted to the specific natures of the complex relations between sensory descriptors and physical features, in order to propose lists of relevant features of different sizes for different descriptors. In order to obtain more reliable results for selection of relevant physical features, the percolation technique has been applied for combining the fuzzy global relevancy and OWA global relevancy criteria in order to clearly distinguish scores of the relevant physical features from those of irrelevant ones.Keywords: data sensitivity, feature selection, fuzzy logic, OWA operators, percolation technique
Procedia PDF Downloads 6054934 A Comprehensive Review of Foam Assisted Water Alternating Gas (FAWAG) Technique: Foam Applications and Mechanisms
Authors: A. Shabib-Asl, M. Abdalla Ayoub Mohammed, A. F. Alta’ee, I. Bin Mohd Saaid, P. Paulo Jose Valentim
Abstract:
In the last few decades, much focus has been placed on enhancing oil recovery from existing fields. This is accomplished by the study and application of various methods. As for recent cases, the study of fluid mobility control and sweep efficiency in gas injection process as well as water alternating gas (WAG) method have demonstrated positive results on oil recovery and thus gained wide interest in petroleum industry. WAG injection application results in an increased oil recovery. Its mechanism consists in reduction of gas oil ratio (GOR). However, there are some problems associated with this which includes poor volumetric sweep efficiency due to its low density and high mobility when compared with oil. This has led to the introduction of foam assisted water alternating gas (FAWAG) technique, which in contrast with WAG injection, acts in improving the sweep efficiency and reducing the gas oil ration therefore maximizing the production rate from the producer wells. This paper presents a comprehensive review of FAWAG process from perspective of Snorre field experience. In addition, some comparative results between FAWAG and the other EOR methods are presented including their setbacks. The main aim is to provide a solid background for future laboratory research and successful field application-extend.Keywords: GOR, mobility ratio, sweep efficiency, WAG
Procedia PDF Downloads 4534933 Multi-Temporal Mapping of Built-up Areas Using Daytime and Nighttime Satellite Images Based on Google Earth Engine Platform
Authors: S. Hutasavi, D. Chen
Abstract:
The built-up area is a significant proxy to measure regional economic growth and reflects the Gross Provincial Product (GPP). However, an up-to-date and reliable database of built-up areas is not always available, especially in developing countries. The cloud-based geospatial analysis platform such as Google Earth Engine (GEE) provides an opportunity with accessibility and computational power for those countries to generate the built-up data. Therefore, this study aims to extract the built-up areas in Eastern Economic Corridor (EEC), Thailand using day and nighttime satellite imagery based on GEE facilities. The normalized indices were generated from Landsat 8 surface reflectance dataset, including Normalized Difference Built-up Index (NDBI), Built-up Index (BUI), and Modified Built-up Index (MBUI). These indices were applied to identify built-up areas in EEC. The result shows that MBUI performs better than BUI and NDBI, with the highest accuracy of 0.85 and Kappa of 0.82. Moreover, the overall accuracy of classification was improved from 79% to 90%, and error of total built-up area was decreased from 29% to 0.7%, after night-time light data from the Visible and Infrared Imaging Suite (VIIRS) Day Night Band (DNB). The results suggest that MBUI with night-time light imagery is appropriate for built-up area extraction and be utilize for further study of socioeconomic impacts of regional development policy over the EEC region.Keywords: built-up area extraction, google earth engine, adaptive thresholding method, rapid mapping
Procedia PDF Downloads 1264932 Recurrent Wheezing and Associated Factors among 6-Year-Old Children in Adama Comprehensive Specialized Hospital Medical College
Authors: Samrawit Tamrat Gebretsadik
Abstract:
Recurrent wheezing is a common respiratory symptom among children, often indicative of underlying airway inflammation and hyperreactivity. Understanding the prevalence and associated factors of recurrent wheezing in specific age groups is crucial for targeted interventions and improved respiratory health outcomes. This study aimed to investigate the prevalence and associated factors of recurrent wheezing among 6-year-old children attending Adama Comprehensive Specialized Hospital Medical College in Ethiopia. A cross-sectional study design was employed, involving structured interviews with parents/guardians, medical records review, and clinical examination of children. Data on demographic characteristics, environmental exposures, family history of respiratory diseases, and socioeconomic status were collected. Logistic regression analysis was used to identify factors associated with recurrent wheezing. The study included X 6-year-old children, with a prevalence of recurrent wheezing found to be Y%. Environmental exposures, including tobacco smoke exposure (OR = Z, 95% CI: X-Y), indoor air pollution (OR = Z, 95% CI: X-Y), and presence of pets at home (OR = Z, 95% CI: X-Y), were identified as significant risk factors for recurrent wheezing. Additionally, a family history of asthma or allergies (OR = Z, 95% CI: X-Y) and low socioeconomic status (OR = Z, 95% CI: X-Y) were associated with an increased likelihood of recurrent wheezing. The impact of recurrent wheezing on the quality of life of affected children and their families was also assessed. Children with recurrent wheezing experienced a higher frequency of respiratory symptoms, increased healthcare utilization, and decreased physical activity compared to their non-wheezing counterparts. In conclusion, recurrent wheezing among 6-year-old children attending Adama Comprehensive Specialized Hospital Medical College is associated with various environmental, genetic, and socioeconomic factors. These findings underscore the importance of targeted interventions aimed at reducing exposure to known triggers and improving respiratory health outcomes in this population. Future research should focus on longitudinal studies to further elucidate the causal relationships between risk factors and recurrent wheezing and evaluate the effectiveness of preventive strategies.Keywords: wheezing, inflammation, respiratory, crucial
Procedia PDF Downloads 534931 The Development of Leisure and Endowment Characteristic Villages in the Perspective of Balancing the Dwellers and Aged Visitors:A Case Study of Villages in Hangzhou Metropolitan Area
Authors: Zijiao Chai, Wangming Li
Abstract:
Under the background of increasing aging population, the situation of city endowment resources shortage gradually revealed. And many villages in the metropolitan area with the good natural ecological environment and leisure tourism base, have become one of the main destinations of urban old people for the off-site pension. This paper is based on a survey of more than ten villages which are characterized by leisure and endowment in Hangzhou metropolitan area, China. The satisfaction degree of the two main groups in the villages, dwellers, and aged visitors, is researched using the method of fuzzy comprehensive evaluation. The statistics are obtained from 535 questionnaires and qualitative interview. According to the satisfaction scores, it could be determined whether the dwellers and aged visitors have reached the equilibrium state. The equilibrium state is the development target of the villages, and it`s defined by environmentally friendly, proper for employment and pension, facilities sharing and harmonious life for each other. Furthermore, this paper comes up with some planning countermeasures in order to avoid "imbalance between dwellers and aged visitors" and obtain sustainable development while maintaining the economic benefit.Keywords: aged visitors, balance between dwellers and aged visitors, dwellers, fuzzy comprehensive evaluation, Hangzhou metropolitan area, leisure and endowment characteristic villages
Procedia PDF Downloads 2894930 Analysis Of Fine Motor Skills in Chronic Neurodegenerative Models of Huntington’s Disease and Amyotrophic Lateral Sclerosis
Authors: T. Heikkinen, J. Oksman, T. Bragge, A. Nurmi, O. Kontkanen, T. Ahtoniemi
Abstract:
Motor impairment is an inherent phenotypic feature of several chronic neurodegenerative diseases, and pharmacological therapies aimed to counterbalance the motor disability have a great market potential. Animal models of chronic neurodegenerative diseases display a number deteriorating motor phenotype during the disease progression. There is a wide array of behavioral tools to evaluate motor functions in rodents. However, currently existing methods to study motor functions in rodents are often limited to evaluate gross motor functions only at advanced stages of the disease phenotype. The most commonly applied traditional motor assays used in CNS rodent models, lack the sensitivity to capture fine motor impairments or improvements. Fine motor skill characterization in rodents provides a more sensitive tool to capture more subtle motor dysfunctions and therapeutic effects. Importantly, similar approach, kinematic movement analysis, is also used in clinic, and applied both in diagnosis and determination of therapeutic response to pharmacological interventions. The aim of this study was to apply kinematic gait analysis, a novel and automated high precision movement analysis system, to characterize phenotypic deficits in three different chronic neurodegenerative animal models, a transgenic mouse model (SOD1 G93A) for amyotrophic lateral sclerosis (ALS), and R6/2 and Q175KI mouse models for Huntington’s disease (HD). The readouts from walking behavior included gait properties with kinematic data, and body movement trajectories including analysis of various points of interest such as movement and position of landmarks in the torso, tail and joints. Mice (transgenic and wild-type) from each model were analyzed for the fine motor kinematic properties at young ages, prior to the age when gross motor deficits are clearly pronounced. Fine motor kinematic Evaluation was continued in the same animals until clear motor dysfunction with conventional motor assays was evident. Time course analysis revealed clear fine motor skill impairments in each transgenic model earlier than what is seen with conventional gross motor tests. Motor changes were quantitatively analyzed for up to ~80 parameters, and the largest data sets of HD models were further processed with principal component analysis (PCA) to transform the pool of individual parameters into a smaller and focused set of mutually uncorrelated gait parameters showing strong genotype difference. Kinematic fine motor analysis of transgenic animal models described in this presentation show that this method isa sensitive, objective and fully automated tool that allows earlier and more sensitive detection of progressive neuromuscular and CNS disease phenotypes. As a result of the analysis a comprehensive set of fine motor parameters for each model is created, and these parameters provide better understanding of the disease progression and enhanced sensitivity of this assay for therapeutic testing compared to classical motor behavior tests. In SOD1 G93A, R6/2, and Q175KI mice, the alterations in gait were evident already several weeks earlier than with traditional gross motor assays. Kinematic testing can be applied to a wider set of motor readouts beyond gait in order to study whole body movement patterns such as with relation to joints and various body parts longitudinally, providing a sophisticated and translatable method for disseminating motor components in rodent disease models and evaluating therapeutic interventions.Keywords: Gait analysis, kinematic, motor impairment, inherent feature
Procedia PDF Downloads 3554929 Computational Model of Human Cardiopulmonary System
Authors: Julian Thrash, Douglas Folk, Michael Ciracy, Audrey C. Tseng, Kristen M. Stromsodt, Amber Younggren, Christopher Maciolek
Abstract:
The cardiopulmonary system is comprised of the heart, lungs, and many dynamic feedback mechanisms that control its function based on a multitude of variables. The next generation of cardiopulmonary medical devices will involve adaptive control and smart pacing techniques. However, testing these smart devices on living systems may be unethical and exceedingly expensive. As a solution, a comprehensive computational model of the cardiopulmonary system was implemented in Simulink. The model contains over 240 state variables and over 100 equations previously described in a series of published articles. Simulink was chosen because of its ease of introducing machine learning elements. Initial results indicate that physiologically correct waveforms of pressures and volumes were obtained in the simulation. With the development of a comprehensive computational model, we hope to pioneer the future of predictive medicine by applying our research towards the initial stages of smart devices. After validation, we will introduce and train reinforcement learning agents using the cardiopulmonary model to assist in adaptive control system design. With our cardiopulmonary model, we will accelerate the design and testing of smart and adaptive medical devices to better serve those with cardiovascular disease.Keywords: adaptive control, cardiopulmonary, computational model, machine learning, predictive medicine
Procedia PDF Downloads 1814928 Effect of Extraction Methods on the Fatty Acids and Physicochemical Properties of Serendipity Berry Seed Oil
Authors: Olufunmilola A. Abiodun, Adegbola O. Dauda, Ayobami Ojo, Samson A. Oyeyinka
Abstract:
Serendipity berry (Dioscoreophyllum cumminsii diel) is a tropical dioecious rainforest vine and native to tropical Africa. The vine grows during the raining season and is used mainly as sweetener. The sweetener in the berry is known as monellin which is sweeter than sucrose. The sweetener is extracted from the fruits and the seed is discarded. The discarded seeds contain bitter principles but had high yield of oil. Serendipity oil was extracted using three methods (N-hexane, expression and expression/n-hexane). Fatty acids and physicochemical properties of the oil obtained were determined. The oil obtained was clear, liquid and have odour similar to hydrocarbon. The percentage oil yield was 38.59, 12.34 and 49.57% for hexane, expression and expression-hexane method respectively. The seed contained high percentage of oil especially using combination of expression and hexane. Low percentage of oil was obtained using expression method. The refractive index values obtained were 1.443, 1.442 and 1.478 for hexane, expression and expression-hexane methods respectively. Peroxide value obtained for expression-hexane was higher than those for hexane and expression. The viscosities of the oil were 125.8, 128.76 and 126.87 cm³/s for hexane, expression and expression-hexane methods respectively which showed that the oil from expression method was more viscous than the other oils. The major fatty acids in serendipity seed oil were oleic acid (62.81%), linoleic acid (22.65%), linolenic (6.11%), palmitic acid (5.67%), stearic acid (2.21%) in decreasing order. Oleic acid which is monounsaturated fatty acid had the highest value. Total unsaturated fatty acids were 91.574, 92.256 and 90.426% for hexane, expression, and expression-hexane respectively. Combination of expression and hexane for extraction of serendipity oil produced high yield of oil. The oil could be refined for food and non-food application.Keywords: serendipity seed oil, expression method, fatty acid, hexane
Procedia PDF Downloads 2734927 Comprehensive Review of Ultralightweight Security Protocols
Authors: Prashansa Singh, Manjot Kaur, Rohit Bajaj
Abstract:
The proliferation of wireless sensor networks and Internet of Things (IoT) devices in the quickly changing digital landscape has highlighted the urgent need for strong security solutions that can handle these systems’ limited resources. A key solution to this problem is the emergence of ultralightweight security protocols, which provide strong security features while respecting the strict computational, energy, and memory constraints imposed on these kinds of devices. This in-depth analysis explores the field of ultralightweight security protocols, offering a thorough examination of their evolution, salient features, and the particular security issues they resolve. We carefully examine and contrast different protocols, pointing out their advantages and disadvantages as well as the compromises between resource limitations and security resilience. We also study these protocols’ application domains, including the Internet of Things, RFID systems, and wireless sensor networks, to name a few. In addition, the review highlights recent developments and advancements in the field, pointing out new trends and possible avenues for future research. This paper aims to be a useful resource for researchers, practitioners, and developers, guiding the design and implementation of safe, effective, and scalable systems in the Internet of Things era by providing a comprehensive overview of ultralightweight security protocols.Keywords: wireless sensor network, machine-to-machine, MQTT broker, server, ultralightweight, TCP/IP
Procedia PDF Downloads 824926 Prevalence and Risk Factors of Low Back Disorder among Waste Collection Workers: A Systematic Review
Authors: Benedicta Asante, Catherine Trask, Brenna Bath
Abstract:
Background: Waste Collection Workers’ (WCWs) activities contribute greatly to the recycling sector and are an important component of the waste management industry. As the recycling sector evolves, there is the increase in reports of injuries, particularly for common and debilitating musculoskeletal disorders such as low back disorder (LBD). WCWs are likely exposed to diverse work-related hazards that could contribute to LBD. However, there is currently no summary of the state of knowledge on the prevalence and risk factors of LBD within this workforce. Method: A comprehensive search was conducted in Ovid Medline, EMBASE, and Global Health e-publications with search term categories ‘low back disorder’ and ‘waste collection workers’. Two reviewers screened articles at title, abstract, and full-text stages. Data were extracted on study design, sampling strategy, socio-demographics, geographical region, and exposure definition, the definition of LBD, response rate, statistical techniques, LBD prevalence and risk factors. The risk of bias was assessed with a standardized tool. Results: The search of three databases generated 79 studies. Thirty-two studies met the study inclusion criteria for both title and abstract; only thirteen full-text articles met the study criteria and underwent data extraction. The majority of articles reported a 12-month prevalence of LBD between 16-74%. Although none of the included studies quantified relationships between risk factors and LBD, the suggested risk factors for LBD among WCWs included: awkward posture; lifting; pulling; pushing; repetitive motions; work duration; and physical loads. Conclusion: LBD is a major occupational health issue among WCWs. In light of these risks and future growth in this industry, further research should focus on the investigation of risk factors, with more focus on ergonomic exposure assessment, and LBD prevention efforts.Keywords: low back pain, scavenger, waste pickers, waste collection workers
Procedia PDF Downloads 2544925 Automatic Differential Diagnosis of Melanocytic Skin Tumours Using Ultrasound and Spectrophotometric Data
Authors: Kristina Sakalauskiene, Renaldas Raisutis, Gintare Linkeviciute, Skaidra Valiukeviciene
Abstract:
Cutaneous melanoma is a melanocytic skin tumour, which has a very poor prognosis while is highly resistant to treatment and tends to metastasize. Thickness of melanoma is one of the most important biomarker for stage of disease, prognosis and surgery planning. In this study, we hypothesized that the automatic analysis of spectrophotometric images and high-frequency ultrasonic 2D data can improve differential diagnosis of cutaneous melanoma and provide additional information about tumour penetration depth. This paper presents the novel complex automatic system for non-invasive melanocytic skin tumour differential diagnosis and penetration depth evaluation. The system is composed of region of interest segmentation in spectrophotometric images and high-frequency ultrasound data, quantitative parameter evaluation, informative feature extraction and classification with linear regression classifier. The segmentation of melanocytic skin tumour region in ultrasound image is based on parametric integrated backscattering coefficient calculation. The segmentation of optical image is based on Otsu thresholding. In total 29 quantitative tissue characterization parameters were evaluated by using ultrasound data (11 acoustical, 4 shape and 15 textural parameters) and 55 quantitative features of dermatoscopic and spectrophotometric images (using total melanin, dermal melanin, blood and collagen SIAgraphs acquired using spectrophotometric imaging device SIAscope). In total 102 melanocytic skin lesions (including 43 cutaneous melanomas) were examined by using SIAscope and ultrasound system with 22 MHz center frequency single element transducer. The diagnosis and Breslow thickness (pT) of each MST were evaluated during routine histological examination after excision and used as a reference. The results of this study have shown that automatic analysis of spectrophotometric and high frequency ultrasound data can improve non-invasive classification accuracy of early-stage cutaneous melanoma and provide supplementary information about tumour penetration depth.Keywords: cutaneous melanoma, differential diagnosis, high-frequency ultrasound, melanocytic skin tumours, spectrophotometric imaging
Procedia PDF Downloads 2704924 Object-Scene: Deep Convolutional Representation for Scene Classification
Authors: Yanjun Chen, Chuanping Hu, Jie Shao, Lin Mei, Chongyang Zhang
Abstract:
Traditional image classification is based on encoding scheme (e.g. Fisher Vector, Vector of Locally Aggregated Descriptor) with low-level image features (e.g. SIFT, HoG). Compared to these low-level local features, deep convolutional features obtained at the mid-level layer of convolutional neural networks (CNN) have richer information but lack of geometric invariance. For scene classification, there are scattered objects with different size, category, layout, number and so on. It is crucial to find the distinctive objects in scene as well as their co-occurrence relationship. In this paper, we propose a method to take advantage of both deep convolutional features and the traditional encoding scheme while taking object-centric and scene-centric information into consideration. First, to exploit the object-centric and scene-centric information, two CNNs that trained on ImageNet and Places dataset separately are used as the pre-trained models to extract deep convolutional features at multiple scales. This produces dense local activations. By analyzing the performance of different CNNs at multiple scales, it is found that each CNN works better in different scale ranges. A scale-wise CNN adaption is reasonable since objects in scene are at its own specific scale. Second, a fisher kernel is applied to aggregate a global representation at each scale and then to merge into a single vector by using a post-processing method called scale-wise normalization. The essence of Fisher Vector lies on the accumulation of the first and second order differences. Hence, the scale-wise normalization followed by average pooling would balance the influence of each scale since different amount of features are extracted. Third, the Fisher vector representation based on the deep convolutional features is followed by a linear Supported Vector Machine, which is a simple yet efficient way to classify the scene categories. Experimental results show that the scale-specific feature extraction and normalization with CNNs trained on object-centric and scene-centric datasets can boost the results from 74.03% up to 79.43% on MIT Indoor67 when only two scales are used (compared to results at single scale). The result is comparable to state-of-art performance which proves that the representation can be applied to other visual recognition tasks.Keywords: deep convolutional features, Fisher Vector, multiple scales, scale-specific normalization
Procedia PDF Downloads 3314923 A Robust and Efficient Segmentation Method Applied for Cardiac Left Ventricle with Abnormal Shapes
Authors: Peifei Zhu, Zisheng Li, Yasuki Kakishita, Mayumi Suzuki, Tomoaki Chono
Abstract:
Segmentation of left ventricle (LV) from cardiac ultrasound images provides a quantitative functional analysis of the heart to diagnose disease. Active Shape Model (ASM) is a widely used approach for LV segmentation but suffers from the drawback that initialization of the shape model is not sufficiently close to the target, especially when dealing with abnormal shapes in disease. In this work, a two-step framework is proposed to improve the accuracy and speed of the model-based segmentation. Firstly, a robust and efficient detector based on Hough forest is proposed to localize cardiac feature points, and such points are used to predict the initial fitting of the LV shape model. Secondly, to achieve more accurate and detailed segmentation, ASM is applied to further fit the LV shape model to the cardiac ultrasound image. The performance of the proposed method is evaluated on a dataset of 800 cardiac ultrasound images that are mostly of abnormal shapes. The proposed method is compared to several combinations of ASM and existing initialization methods. The experiment results demonstrate that the accuracy of feature point detection for initialization was improved by 40% compared to the existing methods. Moreover, the proposed method significantly reduces the number of necessary ASM fitting loops, thus speeding up the whole segmentation process. Therefore, the proposed method is able to achieve more accurate and efficient segmentation results and is applicable to unusual shapes of heart with cardiac diseases, such as left atrial enlargement.Keywords: hough forest, active shape model, segmentation, cardiac left ventricle
Procedia PDF Downloads 3394922 Global Voltage Harmonic Index for Measuring Harmonic Situation of Power Grids: A Focus on Power Transformers
Authors: Alireza Zabihi, Saeed Peyghami, Hossein Mokhtari
Abstract:
With the increasing deployment of renewable power plants, such as solar and wind, it is crucial to measure the harmonic situation of the grid. This paper proposes a global voltage harmonic index to measure the harmonic situation of the power grid with a focus on power transformers. The power electronics systems used to connect these plants to the network can introduce harmonics, leading to increased losses, reduced efficiency, false operation of protective relays, and equipment damage due to harmonic intensifications. The proposed index considers the losses caused by harmonics in power transformers which are of great importance and value to the network, providing a comprehensive measure of the harmonic situation of the grid. The effectiveness of the proposed index is evaluated on a real-world distribution network, and the results demonstrate its ability to identify the harmonic situation of the network, particularly in relation to power transformers. The proposed index provides a comprehensive measure of the harmonic situation of the grid, taking into account the losses caused by harmonics in power transformers. The proposed index has the potential to support power companies in optimizing their power systems and to guide researchers in developing effective mitigation strategies for harmonics in the power grid.Keywords: global voltage harmonic index, harmonics, power grid, power quality, power transformers, renewable energy
Procedia PDF Downloads 1274921 Revitalization Strategy of Beijing-Tianjin-Hebei Rural Areas Organized by Production-Living-Ecology Spatial Network at Township Level
Authors: Liuhui Zhu, Peng Zeng
Abstract:
The rural revitalization strategy means to take the country and the city on the same level, and achieve urban-rural integration and comprehensive development of rural areas. Beijing-Tianjin-Hebei rural areas have always been the weak links in the region, with prominently uneven development between urban and rural areas. The rural areas need to join the overall regional synergy. Based on the analysis of the characteristics and problems of rural development in the region from the perspective of production-living-ecology space, the paper proposes the township as the basic unit for rural revitalization according to the overall requirements of the rural revitalization strategy. The basic unit helps to realize resource arrangement, functional organization, and collaborative governance organized by the production-living-ecology spatial network. The paper summarizes the planning strategies for the basic unit. Through spatial cognition and spatial reconstruction, the three space is networked through the base, nodes, and connections to improve the comprehensive value of rural areas and achieve the multiple goals of rural revitalization.Keywords: rural revitalization, Beijing-Tianjin-Hebei region, township level, production-living-ecology spatial network
Procedia PDF Downloads 1954920 Rapid Identification and Diagnosis of the Pathogenic Leptospiras through Comparison among Culture, PCR and Real Time PCR Techniques from Samples of Human and Mouse Feces
Authors: S. Rostampour Yasouri, M. Ghane, M. Doudi
Abstract:
Leptospirosis is one of the most significant infectious and zoonotic diseases along with global spreading. This disease is causative agent of economoic losses and human fatalities in various countries, including Northern provinces of Iran. The aim of this research is to identify and compare the rapid diagnostic techniques of pathogenic leptospiras, considering the multifacetedness of the disease from a clinical manifestation and premature death of patients. In the spring and summer of 2020-2022, 25 fecal samples were collected from suspected leptospirosis patients and 25 Fecal samples from mice residing in the rice fields and factories in Tonekabon city. Samples were prepared by centrifugation and passing through membrane filters. Culture technique was used in liquid and solid EMJH media during one month of incubation at 30°C. Then, the media were examined microscopically. DNA extraction was conducted by extraction Kit. Diagnosis of leptospiras was enforced by PCR and Real time PCR (SYBR Green) techniques using lipL32 specific primer. Out of the patients, 11 samples (44%) and 8 samples (32%) were determined to be pathogenic Leptospira by Real time PCR and PCR technique, respectively. Out of the mice, 9 Samples (36%) and 3 samples (12%) were determined to be pathogenic Leptospira by the mentioned techniques, respectively. Although the culture technique is considered to be the gold standard technique, but due to the slow growth of pathogenic Leptospira and lack of colony formation of some species, it is not a fast technique. Real time PCR allowed rapid diagnosis with much higher accuracy compared to PCR because PCR could not completely identify samples with lower microbial load.Keywords: culture, pathogenic leptospiras, PCR, real time PCR
Procedia PDF Downloads 854919 Recognition and Counting Algorithm for Sub-Regional Objects in a Handwritten Image through Image Sets
Authors: Kothuri Sriraman, Mattupalli Komal Teja
Abstract:
In this paper, a novel algorithm is proposed for the recognition of hulls in a hand written images that might be irregular or digit or character shape. Identification of objects and internal objects is quite difficult to extract, when the structure of the image is having bulk of clusters. The estimation results are easily obtained while going through identifying the sub-regional objects by using the SASK algorithm. Focusing mainly to recognize the number of internal objects exist in a given image, so as it is shadow-free and error-free. The hard clustering and density clustering process of obtained image rough set is used to recognize the differentiated internal objects, if any. In order to find out the internal hull regions it involves three steps pre-processing, Boundary Extraction and finally, apply the Hull Detection system. By detecting the sub-regional hulls it can increase the machine learning capability in detection of characters and it can also be extend in order to get the hull recognition even in irregular shape objects like wise black holes in the space exploration with their intensities. Layered hulls are those having the structured layers inside while it is useful in the Military Services and Traffic to identify the number of vehicles or persons. This proposed SASK algorithm is helpful in making of that kind of identifying the regions and can useful in undergo for the decision process (to clear the traffic, to identify the number of persons in the opponent’s in the war).Keywords: chain code, Hull regions, Hough transform, Hull recognition, Layered Outline Extraction, SASK algorithm
Procedia PDF Downloads 3494918 Code Embedding for Software Vulnerability Discovery Based on Semantic Information
Authors: Joseph Gear, Yue Xu, Ernest Foo, Praveen Gauravaran, Zahra Jadidi, Leonie Simpson
Abstract:
Deep learning methods have been seeing an increasing application to the long-standing security research goal of automatic vulnerability detection for source code. Attention, however, must still be paid to the task of producing vector representations for source code (code embeddings) as input for these deep learning models. Graphical representations of code, most predominantly Abstract Syntax Trees and Code Property Graphs, have received some use in this task of late; however, for very large graphs representing very large code snip- pets, learning becomes prohibitively computationally expensive. This expense may be reduced by intelligently pruning this input to only vulnerability-relevant information; however, little research in this area has been performed. Additionally, most existing work comprehends code based solely on the structure of the graph at the expense of the information contained by the node in the graph. This paper proposes Semantic-enhanced Code Embedding for Vulnerability Discovery (SCEVD), a deep learning model which uses semantic-based feature selection for its vulnerability classification model. It uses information from the nodes as well as the structure of the code graph in order to select features which are most indicative of the presence or absence of vulnerabilities. This model is implemented and experimentally tested using the SARD Juliet vulnerability test suite to determine its efficacy. It is able to improve on existing code graph feature selection methods, as demonstrated by its improved ability to discover vulnerabilities.Keywords: code representation, deep learning, source code semantics, vulnerability discovery
Procedia PDF Downloads 1584917 Uplifting Citizens Participation: A Gov 2.0 Framework
Authors: Mohammed Aladalah
Abstract:
The emergence of digital citizens is no longer mere speculation; therefore, governments’ use of Web 2.0 tools (hereafter Gov 2.0) should be a part of all current and future e-government plans. The potential of Gov 2.0 to facilitate greater communication, participation, and collaboration with citizens has been highlighted and discussed extensively in recent literature. However, the current levels of citizens’ participation in Gov 2.0 have not lived up to the hype. Therefore, governments need to rethink the way in which they implement Gov 2.0, and take advantage of the digitally-engaged population. We propose a two-dimensional framework to tackle this issue: first, on the supply side, governments tend to use Gov 2.0 mainly for the dissemination of information and for self-promotion without the desire to encourage any interaction with citizens; this is due to many reasons, including the lack of time and the possibility of loss of control. The second dimension of the framework is the demand side; citizens are unwilling to participate in Gov 2.0 activities because they do not perceive its value or trust the government. We attempt to consider the elements of both supply and demand in order to provide a comprehensive solution whereby the potential of Gov 2.0 can be fully utilized. Our framework is based on the theoretical foundation of service science and value co-creation theory. This paper makes two significant contributions: (a) it provides an initial framework intended to increase citizens’ participation in Gov 2.0; and (b) it enhances the understanding of the government’s Gov 2.0 applications, particularly in terms of factors that ensure their attractiveness for citizens. This work is the first step in a comprehensive research undertaking, the purpose of which is to study public’s engagement with the Gov 2.0 concept. It contributes to providing a better understanding of e-government and its future.Keywords: e-government, Gov 2.0, citizens participation, digital citizen
Procedia PDF Downloads 3364916 Design of Evaluation for Ehealth Intervention: A Participatory Study in Italy, Israel, Spain and Sweden
Authors: Monika Jurkeviciute, Amia Enam, Johanna Torres Bonilla, Henrik Eriksson
Abstract:
Introduction: Many evaluations of eHealth interventions conclude that the evidence for improved clinical outcomes is limited, especially when the intervention is short, such as one year. Often, evaluation design does not address the feasibility of achieving clinical outcomes. Evaluations are designed to reflect upon clinical goals of intervention without utilizing the opportunity to illuminate effects on organizations and cost. A comprehensive design of evaluation can better support decision-making regarding the effectiveness and potential transferability of eHealth. Hence, the purpose of this paper is to present a feasible and comprehensive design of evaluation for eHealth intervention, including the design process in different contexts. Methodology: The situation of limited feasibility of clinical outcomes was foreseen in the European Union funded project called “DECI” (“Digital Environment for Cognitive Inclusion”) that is run under the “Horizon 2020” program with an aim to define and test a digital environment platform within corresponding care models that help elderly people live independently. A complex intervention of eHealth implementation into elaborate care models in four different countries was planned for one year. To design the evaluation, a participative approach was undertaken using Pettigrew’s lens of change and transformations, including context, process, and content. Through a series of workshops, observations, interviews, and document analysis, as well as a review of scientific literature, a comprehensive design of evaluation was created. Findings: The findings indicate that in order to get evidence on clinical outcomes, eHealth interventions should last longer than one year. The content of the comprehensive evaluation design includes a collection of qualitative and quantitative methods for data gathering which illuminates non-medical aspects. Furthermore, it contains communication arrangements to discuss the results and continuously improve the evaluation design, as well as procedures for monitoring and improving the data collection during the intervention. The process of the comprehensive evaluation design consists of four stages: (1) analysis of a current state in different contexts, including measurement systems, expectations and profiles of stakeholders, organizational ambitions to change due to eHealth integration, and the organizational capacity to collect data for evaluation; (2) workshop with project partners to discuss the as-is situation in relation to the project goals; (3) development of general and customized sets of relevant performance measures, questionnaires and interview questions; (4) setting up procedures and monitoring systems for the interventions. Lastly, strategies are presented on how challenges can be handled during the design process of evaluation in four different countries. The evaluation design needs to consider contextual factors such as project limitations, and differences between pilot sites in terms of eHealth solutions, patient groups, care models, national and organizational cultures and settings. This implies a need for the flexible approach to evaluation design to enable judgment over the effectiveness and potential for adoption and transferability of eHealth. In summary, this paper provides learning opportunities for future evaluation designs of eHealth interventions in different national and organizational settings.Keywords: ehealth, elderly, evaluation, intervention, multi-cultural
Procedia PDF Downloads 3244915 Uncertainty Quantification of Corrosion Anomaly Length of Oil and Gas Steel Pipelines Based on Inline Inspection and Field Data
Authors: Tammeen Siraj, Wenxing Zhou, Terry Huang, Mohammad Al-Amin
Abstract:
The high resolution inline inspection (ILI) tool is used extensively in the pipeline industry to identify, locate, and measure metal-loss corrosion anomalies on buried oil and gas steel pipelines. Corrosion anomalies may occur singly (i.e. individual anomalies) or as clusters (i.e. a colony of corrosion anomalies). Although the ILI technology has advanced immensely, there are measurement errors associated with the sizes of corrosion anomalies reported by ILI tools due limitations of the tools and associated sizing algorithms, and detection threshold of the tools (i.e. the minimum detectable feature dimension). Quantifying the measurement error in the ILI data is crucial for corrosion management and developing maintenance strategies that satisfy the safety and economic constraints. Studies on the measurement error associated with the length of the corrosion anomalies (in the longitudinal direction of the pipeline) has been scarcely reported in the literature and will be investigated in the present study. Limitations in the ILI tool and clustering process can sometimes cause clustering error, which is defined as the error introduced during the clustering process by including or excluding a single or group of anomalies in or from a cluster. Clustering error has been found to be one of the biggest contributory factors for relatively high uncertainties associated with ILI reported anomaly length. As such, this study focuses on developing a consistent and comprehensive framework to quantify the measurement errors in the ILI-reported anomaly length by comparing the ILI data and corresponding field measurements for individual and clustered corrosion anomalies. The analysis carried out in this study is based on the ILI and field measurement data for a set of anomalies collected from two segments of a buried natural gas pipeline currently in service in Alberta, Canada. Data analyses showed that the measurement error associated with the ILI-reported length of the anomalies without clustering error, denoted as Type I anomalies is markedly less than that for anomalies with clustering error, denoted as Type II anomalies. A methodology employing data mining techniques is further proposed to classify the Type I and Type II anomalies based on the ILI-reported corrosion anomaly information.Keywords: clustered corrosion anomaly, corrosion anomaly assessment, corrosion anomaly length, individual corrosion anomaly, metal-loss corrosion, oil and gas steel pipeline
Procedia PDF Downloads 3094914 The Long-Term Effects of Immediate Implantation, Early Implantation and Delayed Implantation at Aesthetics Area
Authors: Xing Wang, Lin Feng, Xuan Zou, Hongchen liu
Abstract:
Immediate Implantation after tooth extraction is considered to be the ideal way to retain the alveolar bone, but some scholars believe the aesthetic effect in the Early Implantation case are more reliable. In this study, 89 patients were added to this retrospective study up to 5 years. Assessment indicators was including the survival of the implant (peri-implant infection, implant loosening, shedding, crowns and occlusal), aesthetics (color and fullness gums, papilla height, probing depth, X-ray alveolar crest height, the patient's own aesthetic satisfaction, doctors aesthetics score), repair defects around the implant (peri-implant bone changes in height and thickness, whether the use of autologous bone graft, whether to use absorption/repair manual nonabsorbable material), treatment time, cost and the use of antibiotics.The results demonstrated that there is no significant difference in long-term success rate of immediate implantation, early implantation and delayed implantation (p> 0.05). But the results indicated immediate implantation group could get get better aesthetic results after two years (p< 0.05), but may increase the risk of complications and failures (p< 0.05). High-risk indicators include gingival recession, labial bone wall damage, thin gingival biotypes, planting position and occlusal restoration bad and so on. No matter which type of implanting methods was selected, the extraction methods and bone defect amplification techniques are observed as a significant factors on aesthetic effect (p< 0.05).Keywords: immediate implantation, long-term effects, aesthetics area, dental implants
Procedia PDF Downloads 3564913 In Vitro Antioxidant and Cytotoxic Activities Against Human Oral Cancer and Human Laryngeal Cancer of Limonia acidissima L. Bark Extracts
Authors: Kriyapa lairungruang, Arunporn Itharat
Abstract:
Limonia acidissima L. (LA) (Common name: wood apple, Thai name: ma-khwit) is a medicinal plant which has long been used in Thai traditional medicine. Its bark is used for treatment of diarrhea, abscess, wound healing and inflammation and it is also used in oral cancer. Thus, this research aimed to investigate antioxidant and cytotoxic activities of the LA bark extracts produced by various extraction methods. Different extraction procedures were used to extract LA bark for biological activity testing: boiling in water, maceration with 95% ethanol, maceration with 50% ethanol and water boiling of each the 95% and the 50% ethanolic residues. All extracts were tested for antioxidant activity using DPPH radical scavenging assay, cytotoxic activity against human laryngeal epidermoid carcinoma (HEp-2) cells and human oral epidermoid carcinoma (KB) cells using sulforhodamine B (SRB) assay. The results found that the 95% ethanolic extract of LA bark showed the highest antioxidant activity with EC50 values of 29.76±1.88 µg/ml. For cytotoxic activity, the 50% ethanolic extract showed the best cytotoxic activity against HEp-2 and KB cells with IC50 values of 9.55±1.68 and 18.90±0.86 µg/ml, respectively. This study demonstrated that the 95% ethanolic extract of LA bark showed moderate antioxidant activity and the 50% ethanolic extract provided potent cytotoxic activity against HEp-2 and KB cells. These results confirm the traditional use of LA for the treatment of oral cancer and laryngeal cancer, and also support its ongoing use.Keywords: antioxidant activity, cytotoxic activity, Laryngeal epidermoid carcinoma, Limonia acidissima L., oral epidermoid carcinoma
Procedia PDF Downloads 4784912 A Method for Identifying Unusual Transactions in E-commerce Through Extended Data Flow Conformance Checking
Authors: Handie Pramana Putra, Ani Dijah Rahajoe
Abstract:
The proliferation of smart devices and advancements in mobile communication technologies have permeated various facets of life with the widespread influence of e-commerce. Detecting abnormal transactions holds paramount significance in this realm due to the potential for substantial financial losses. Moreover, the fusion of data flow and control flow assumes a critical role in the exploration of process modeling and data analysis, contributing significantly to the accuracy and security of business processes. This paper introduces an alternative approach to identify abnormal transactions through a model that integrates both data and control flows. Referred to as the Extended Data Petri net (DPNE), our model encapsulates the entire process, encompassing user login to the e-commerce platform and concluding with the payment stage, including the mobile transaction process. We scrutinize the model's structure, formulate an algorithm for detecting anomalies in pertinent data, and elucidate the rationale and efficacy of the comprehensive system model. A case study validates the responsive performance of each system component, demonstrating the system's adeptness in evaluating every activity within mobile transactions. Ultimately, the results of anomaly detection are derived through a thorough and comprehensive analysis.Keywords: database, data analysis, DPNE, extended data flow, e-commerce
Procedia PDF Downloads 564911 Hybrid Approach for Face Recognition Combining Gabor Wavelet and Linear Discriminant Analysis
Authors: A: Annis Fathima, V. Vaidehi, S. Ajitha
Abstract:
Face recognition system finds many applications in surveillance and human computer interaction systems. As the applications using face recognition systems are of much importance and demand more accuracy, more robustness in the face recognition system is expected with less computation time. In this paper, a hybrid approach for face recognition combining Gabor Wavelet and Linear Discriminant Analysis (HGWLDA) is proposed. The normalized input grayscale image is approximated and reduced in dimension to lower the processing overhead for Gabor filters. This image is convolved with bank of Gabor filters with varying scales and orientations. LDA, a subspace analysis techniques are used to reduce the intra-class space and maximize the inter-class space. The techniques used are 2-dimensional Linear Discriminant Analysis (2D-LDA), 2-dimensional bidirectional LDA ((2D)2LDA), Weighted 2-dimensional bidirectional Linear Discriminant Analysis (Wt (2D)2 LDA). LDA reduces the feature dimension by extracting the features with greater variance. k-Nearest Neighbour (k-NN) classifier is used to classify and recognize the test image by comparing its feature with each of the training set features. The HGWLDA approach is robust against illumination conditions as the Gabor features are illumination invariant. This approach also aims at a better recognition rate using less number of features for varying expressions. The performance of the proposed HGWLDA approaches is evaluated using AT&T database, MIT-India face database and faces94 database. It is found that the proposed HGWLDA approach provides better results than the existing Gabor approach.Keywords: face recognition, Gabor wavelet, LDA, k-NN classifier
Procedia PDF Downloads 4674910 Color Image Compression/Encryption/Contour Extraction using 3L-DWT and SSPCE Method
Authors: Ali A. Ukasha, Majdi F. Elbireki, Mohammad F. Abdullah
Abstract:
Data security needed in data transmission, storage, and communication to ensure the security. This paper is divided into two parts. This work interests with the color image which is decomposed into red, green and blue channels. The blue and green channels are compressed using 3-levels discrete wavelet transform. The Arnold transform uses to changes the locations of red image channel pixels as image scrambling process. Then all these channels are encrypted separately using the key image that has same original size and are generating using private keys and modulo operations. Performing the X-OR and modulo operations between the encrypted channels images for image pixel values change purpose. The extracted contours from color images recovery can be obtained with accepted level of distortion using single step parallel contour extraction (SSPCE) method. Experiments have demonstrated that proposed algorithm can fully encrypt 2D Color images and completely reconstructed without any distortion. Also shown that the analyzed algorithm has extremely large security against some attacks like salt and pepper and Jpeg compression. Its proof that the color images can be protected with a higher security level. The presented method has easy hardware implementation and suitable for multimedia protection in real time applications such as wireless networks and mobile phone services.Keywords: SSPCE method, image compression and salt and peppers attacks, bitplanes decomposition, Arnold transform, color image, wavelet transform, lossless image encryption
Procedia PDF Downloads 5184909 Speech Detection Model Based on Deep Neural Networks Classifier for Speech Emotions Recognition
Authors: Aisultan Shoiynbek, Darkhan Kuanyshbay, Paulo Menezes, Akbayan Bekarystankyzy, Assylbek Mukhametzhanov, Temirlan Shoiynbek
Abstract:
Speech emotion recognition (SER) has received increasing research interest in recent years. It is a common practice to utilize emotional speech collected under controlled conditions recorded by actors imitating and artificially producing emotions in front of a microphone. There are four issues related to that approach: emotions are not natural, meaning that machines are learning to recognize fake emotions; emotions are very limited in quantity and poor in variety of speaking; there is some language dependency in SER; consequently, each time researchers want to start work with SER, they need to find a good emotional database in their language. This paper proposes an approach to create an automatic tool for speech emotion extraction based on facial emotion recognition and describes the sequence of actions involved in the proposed approach. One of the first objectives in the sequence of actions is the speech detection issue. The paper provides a detailed description of the speech detection model based on a fully connected deep neural network for Kazakh and Russian. Despite the high results in speech detection for Kazakh and Russian, the described process is suitable for any language. To investigate the working capacity of the developed model, an analysis of speech detection and extraction from real tasks has been performed.Keywords: deep neural networks, speech detection, speech emotion recognition, Mel-frequency cepstrum coefficients, collecting speech emotion corpus, collecting speech emotion dataset, Kazakh speech dataset
Procedia PDF Downloads 264908 The Influence of Family of Origin on Children: A Comprehensive Model and Implications for Positive Psychology and Psychotherapy
Authors: Meichen He, Xuan Yang
Abstract:
Background: In the field of psychotherapy, the role of the family of origin is of utmost importance. Over the past few decades, both individual-oriented and family-oriented approaches to child therapy have shown moderate success in reducing children's psychological and behavioral issues. Objective: However, in exploring how the family of origin influences individuals, it has been noted that there is a lack of comprehensive measurement indicators and an absence of an exact model to assess the impact of the family of origin on individual development. Therefore, this study aims to develop a model based on a literature review regarding the influence of the family of origin on children. Specifically, it will examine the effects of factors such as education level, economic status, maternal age, family integration, family violence, marital conflict, parental substance abuse, and alcohol consumption on children's self-confidence and life satisfaction. Through this research, we aim to further investigate the impact of the family of origin on children and provide directions for future research in positive psychology and psychotherapy. Methods: This study will employ a literature review methodology to gather and analyze relevant research articles on the influence of the family of origin on children. Subsequently, we will conduct quantitative analyses to establish a comprehensive model explaining how family of origin factors affect children's psychological and behavioral outcomes. Findings: the research has revealed that family of origin factors, including education level, economic status, maternal age, family integration, family violence, marital conflict, parental drug and alcohol consumption, have an impact on children's self-confidence and life satisfaction. These factors can affect children's psychological well-being and happiness through various pathways. Implications: The results of this study will contribute to a better understanding of the influence of the family of origin on children and provide valuable directions for future research in positive psychology and psychotherapy. This research will enhance awareness of children's psychological well-being and lay the foundation for improving psychotherapeutic methods.Keywords: family of origion, positive psychology, developmental psychology, family education, social psychology, educational psychology
Procedia PDF Downloads 165