Search results for: self-regulated learning theory
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10967

Search results for: self-regulated learning theory

5717 Learning from Dendrites: Improving the Point Neuron Model

Authors: Alexander Vandesompele, Joni Dambre

Abstract:

The diversity in dendritic arborization, as first illustrated by Santiago Ramon y Cajal, has always suggested a role for dendrites in the functionality of neurons. In the past decades, thanks to new recording techniques and optical stimulation methods, it has become clear that dendrites are not merely passive electrical components. They are observed to integrate inputs in a non-linear fashion and actively participate in computations. Regardless, in simulations of neural networks dendritic structure and functionality are often overlooked. Especially in a machine learning context, when designing artificial neural networks, point neuron models such as the leaky-integrate-and-fire (LIF) model are dominant. These models mimic the integration of inputs at the neuron soma, and ignore the existence of dendrites. In this work, the LIF point neuron model is extended with a simple form of dendritic computation. This gives the LIF neuron increased capacity to discriminate spatiotemporal input sequences, a dendritic functionality as observed in another study. Simulations of the spiking neurons are performed using the Bindsnet framework. In the common LIF model, incoming synapses are independent. Here, we introduce a dependency between incoming synapses such that the post-synaptic impact of a spike is not only determined by the weight of the synapse, but also by the activity of other synapses. This is a form of short term plasticity where synapses are potentiated or depressed by the preceding activity of neighbouring synapses. This is a straightforward way to prevent inputs from simply summing linearly at the soma. To implement this, each pair of synapses on a neuron is assigned a variable,representing the synaptic relation. This variable determines the magnitude ofthe short term plasticity. These variables can be chosen randomly or, more interestingly, can be learned using a form of Hebbian learning. We use Spike-Time-Dependent-Plasticity (STDP), commonly used to learn synaptic strength magnitudes. If all neurons in a layer receive the same input, they tend to learn the same through STDP. Adding inhibitory connections between the neurons creates a winner-take-all (WTA) network. This causes the different neurons to learn different input sequences. To illustrate the impact of the proposed dendritic mechanism, even without learning, we attach five input neurons to two output neurons. One output neuron isa regular LIF neuron, the other output neuron is a LIF neuron with dendritic relationships. Then, the five input neurons are allowed to fire in a particular order. The membrane potentials are reset and subsequently the five input neurons are fired in the reversed order. As the regular LIF neuron linearly integrates its inputs at the soma, the membrane potential response to both sequences is similar in magnitude. In the other output neuron, due to the dendritic mechanism, the membrane potential response is different for both sequences. Hence, the dendritic mechanism improves the neuron’s capacity for discriminating spa-tiotemporal sequences. Dendritic computations improve LIF neurons even if the relationships between synapses are established randomly. Ideally however, a learning rule is used to improve the dendritic relationships based on input data. It is possible to learn synaptic strength with STDP, to make a neuron more sensitive to its input. Similarly, it is possible to learn dendritic relationships with STDP, to make the neuron more sensitive to spatiotemporal input sequences. Feeding structured data to a WTA network with dendritic computation leads to a significantly higher number of discriminated input patterns. Without the dendritic computation, output neurons are less specific and may, for instance, be activated by a sequence in reverse order.

Keywords: dendritic computation, spiking neural networks, point neuron model

Procedia PDF Downloads 116
5716 Accurate Mass Segmentation Using U-Net Deep Learning Architecture for Improved Cancer Detection

Authors: Ali Hamza

Abstract:

Accurate segmentation of breast ultrasound images is of paramount importance in enhancing the diagnostic capabilities of breast cancer detection. This study presents an approach utilizing the U-Net architecture for segmenting breast ultrasound images aimed at improving the accuracy and reliability of mass identification within the breast tissue. The proposed method encompasses a multi-stage process. Initially, preprocessing techniques are employed to refine image quality and diminish noise interference. Subsequently, the U-Net architecture, a deep learning convolutional neural network (CNN), is employed for pixel-wise segmentation of regions of interest corresponding to potential breast masses. The U-Net's distinctive architecture, characterized by a contracting and expansive pathway, enables accurate boundary delineation and detailed feature extraction. To evaluate the effectiveness of the proposed approach, an extensive dataset of breast ultrasound images is employed, encompassing diverse cases. Quantitative performance metrics such as the Dice coefficient, Jaccard index, sensitivity, specificity, and Hausdorff distance are employed to comprehensively assess the segmentation accuracy. Comparative analyses against traditional segmentation methods showcase the superiority of the U-Net architecture in capturing intricate details and accurately segmenting breast masses. The outcomes of this study emphasize the potential of the U-Net-based segmentation approach in bolstering breast ultrasound image analysis. The method's ability to reliably pinpoint mass boundaries holds promise for aiding radiologists in precise diagnosis and treatment planning. However, further validation and integration within clinical workflows are necessary to ascertain their practical clinical utility and facilitate seamless adoption by healthcare professionals. In conclusion, leveraging the U-Net architecture for breast ultrasound image segmentation showcases a robust framework that can significantly enhance diagnostic accuracy and advance the field of breast cancer detection. This approach represents a pivotal step towards empowering medical professionals with a more potent tool for early and accurate breast cancer diagnosis.

Keywords: mage segmentation, U-Net, deep learning, breast cancer detection, diagnostic accuracy, mass identification, convolutional neural network

Procedia PDF Downloads 66
5715 A Machine Learning Approach for Earthquake Prediction in Various Zones Based on Solar Activity

Authors: Viacheslav Shkuratskyy, Aminu Bello Usman, Michael O’Dea, Saifur Rahman Sabuj

Abstract:

This paper examines relationships between solar activity and earthquakes; it applied machine learning techniques: K-nearest neighbour, support vector regression, random forest regression, and long short-term memory network. Data from the SILSO World Data Center, the NOAA National Center, the GOES satellite, NASA OMNIWeb, and the United States Geological Survey were used for the experiment. The 23rd and 24th solar cycles, daily sunspot number, solar wind velocity, proton density, and proton temperature were all included in the dataset. The study also examined sunspots, solar wind, and solar flares, which all reflect solar activity and earthquake frequency distribution by magnitude and depth. The findings showed that the long short-term memory network model predicts earthquakes more correctly than the other models applied in the study, and solar activity is more likely to affect earthquakes of lower magnitude and shallow depth than earthquakes of magnitude 5.5 or larger with intermediate depth and deep depth.

Keywords: k-nearest neighbour, support vector regression, random forest regression, long short-term memory network, earthquakes, solar activity, sunspot number, solar wind, solar flares

Procedia PDF Downloads 62
5714 Determining Variables in Mathematics Performance According to Gender in Mexican Elementary School

Authors: Nora Gavira Duron, Cinthya Moreda Gonzalez-Ortega, Reyna Susana Garcia Ruiz

Abstract:

This paper objective is to analyze the mathematics performance in the Learning Evaluation National Plan (PLANEA for its Spanish initials: Plan Nacional para la Evaluación de los Aprendizajes), applied to Mexican students who are enrolled in the last elementary-school year over the 2017-2018 academic year. Such test was conducted nationwide in 3,573 schools, using a sample of 108,083 students, whose average in mathematics, on a scale of 0 to 100, was 45.6 points. 75% of the sample analyzed did not reach the sufficiency level (60 points). It should be noted that only 2% got a 90 or higher score result. The performance is analyzed while considering whether there are differences in gender, marginalization level, public or private school enrollment, parents’ academic background, and living-with-parents situation. Likewise, this variable impact (among other variables) on school performance by gender is evaluated, considering multivariate logistic (Logit) regression analysis. The results show there are no significant differences in mathematics performance regarding gender in elementary school; nevertheless, the impact exerted by mothers who studied at least high school is of great relevance for students, particularly for girls. Other determining variables are students’ resilience, their parents’ economic status, and the fact they attend private schools, strengthened by the mother's education.

Keywords: multivariate regression analysis, academic performance, learning evaluation, mathematics result per gender

Procedia PDF Downloads 133
5713 Encephalon-An Implementation of a Handwritten Mathematical Expression Solver

Authors: Shreeyam, Ranjan Kumar Sah, Shivangi

Abstract:

Recognizing and solving handwritten mathematical expressions can be a challenging task, particularly when certain characters are segmented and classified. This project proposes a solution that uses Convolutional Neural Network (CNN) and image processing techniques to accurately solve various types of equations, including arithmetic, quadratic, and trigonometric equations, as well as logical operations like logical AND, OR, NOT, NAND, XOR, and NOR. The proposed solution also provides a graphical solution, allowing users to visualize equations and their solutions. In addition to equation solving, the platform, called CNNCalc, offers a comprehensive learning experience for students. It provides educational content, a quiz platform, and a coding platform for practicing programming skills in different languages like C, Python, and Java. This all-in-one solution makes the learning process engaging and enjoyable for students. The proposed methodology includes horizontal compact projection analysis and survey for segmentation and binarization, as well as connected component analysis and integrated connected component analysis for character classification. The compact projection algorithm compresses the horizontal projections to remove noise and obtain a clearer image, contributing to the accuracy of character segmentation. Experimental results demonstrate the effectiveness of the proposed solution in solving a wide range of mathematical equations. CNNCalc provides a powerful and user-friendly platform for solving equations, learning, and practicing programming skills. With its comprehensive features and accurate results, CNNCalc is poised to revolutionize the way students learn and solve mathematical equations. The platform utilizes a custom-designed Convolutional Neural Network (CNN) with image processing techniques to accurately recognize and classify symbols within handwritten equations. The compact projection algorithm effectively removes noise from horizontal projections, leading to clearer images and improved character segmentation. Experimental results demonstrate the accuracy and effectiveness of the proposed solution in solving a wide range of equations, including arithmetic, quadratic, trigonometric, and logical operations. CNNCalc features a user-friendly interface with a graphical representation of equations being solved, making it an interactive and engaging learning experience for users. The platform also includes tutorials, testing capabilities, and programming features in languages such as C, Python, and Java. Users can track their progress and work towards improving their skills. CNNCalc is poised to revolutionize the way students learn and solve mathematical equations with its comprehensive features and accurate results.

Keywords: AL, ML, hand written equation solver, maths, computer, CNNCalc, convolutional neural networks

Procedia PDF Downloads 101
5712 Cultural Adaptation of Foreign Students in Vienna, A Sociolinguistic Case Study of Iranian Students in Vienna

Authors: Roshanak Nouralian

Abstract:

The primary focus of my Ph.D. dissertation revolves around the interconnection between language and culture, as well as the crucial role that language plays in facilitating communication and fostering integration within the host society for immigrants. This research specifically focuses on Iranian students studying at various universities in Vienna. Throughout this study, I have attempted to examine and analyze their challenges in various life situations in Austria. The broad dimensions of the research question led the research process to apply a constructivist grounded theory strategy. I have also used critical discourse analysis that is in line with constructivist GT's point of view to look closely at the borders, contradictions, and inequalities that came up in the participants' real-life experiences. Data from individual interviews and group discussions have expanded the research trajectory beyond disciplinary boundaries toward a transdisciplinary approach. The research findings indicate how the language policy of the host society leads to the establishment of power relationships and the arousal of a sense of cultural dominance among the research participants. This study investigates the problems experienced by participants in their daily interactions within the host society. Additionally, the results illustrate the development of a dependency relationship between participants and their host society despite linguistic policies that cause a sense of cultural hegemony. Conversely, the obtained data allowed me to examine the participants' language ideologies. The findings of this study show that social linguistics has the potential to go beyond the boundaries of its field. This is possible by using a variety of research strategies and analyzing people's real-life experiences to find out how language affects different parts of their daily lives. Therefore, in this conference, discussing the logic of employing a constructivist GT strategy along with critical discourse analysis (CDA) in this research, I intend to discuss the achieved results.

Keywords: cultural adapttaion, language policy, language ideology, cultural hegemony, transdisciplinary research, constructivist grounded theory, critical discourse analysis

Procedia PDF Downloads 48
5711 Evidence of Half-Metallicity in Cubic PrMnO3 Perovskite

Authors: B. Bouadjemi, S. Bentata, W. Benstaali, A. Abbad

Abstract:

The electronic and magnetic properties of the cubic praseodymium oxides perovskites PrMnO3 were calculated using the density functional theory (DFT) with both generalized gradient approximation (GGA) and GGA+U approaches, where U is on-site Coulomb interaction correction. The results show a half-metallic ferromagnetic ground state for PrMnO3 in GGA+U approached, while semi-metallic ferromagnetic character is observed in GGA. The results obtained, make the cubic PrMnO3 a promising candidate for application in spintronics.

Keywords: first-principles, electronic properties, transition metal, materials science

Procedia PDF Downloads 449
5710 Reconstructability Analysis for Landslide Prediction

Authors: David Percy

Abstract:

Landslides are a geologic phenomenon that affects a large number of inhabited places and are constantly being monitored and studied for the prediction of future occurrences. Reconstructability analysis (RA) is a methodology for extracting informative models from large volumes of data that work exclusively with discrete data. While RA has been used in medical applications and social science extensively, we are introducing it to the spatial sciences through applications like landslide prediction. Since RA works exclusively with discrete data, such as soil classification or bedrock type, working with continuous data, such as porosity, requires that these data are binned for inclusion in the model. RA constructs models of the data which pick out the most informative elements, independent variables (IVs), from each layer that predict the dependent variable (DV), landslide occurrence. Each layer included in the model retains its classification data as a primary encoding of the data. Unlike other machine learning algorithms that force the data into one-hot encoding type of schemes, RA works directly with the data as it is encoded, with the exception of continuous data, which must be binned. The usual physical and derived layers are included in the model, and testing our results against other published methodologies, such as neural networks, yields accuracy that is similar but with the advantage of a completely transparent model. The results of an RA session with a data set are a report on every combination of variables and their probability of landslide events occurring. In this way, every combination of informative state combinations can be examined.

Keywords: reconstructability analysis, machine learning, landslides, raster analysis

Procedia PDF Downloads 48
5709 Development of Advanced Virtual Radiation Detection and Measurement Laboratory (AVR-DML) for Nuclear Science and Engineering Students

Authors: Lily Ranjbar, Haori Yang

Abstract:

Online education has been around for several decades, but the importance of online education became evident after the COVID-19 pandemic. Eventhough the online delivery approach works well for knowledge building through delivering content and oversight processes, it has limitations in developing hands-on laboratory skills, especially in the STEM field. During the pandemic, many education institutions faced numerous challenges in delivering lab-based courses, especially in the STEM field. Also, many students worldwide were unable to practice working with lab equipment due to social distancing or the significant cost of highly specialized equipment. The laboratory plays a crucial role in nuclear science and engineering education. It can engage students and improve their learning outcomes. In addition, online education and virtual labs have gained substantial popularity in engineering and science education. Therefore, developing virtual labs is vital for institutions to deliver high-class education to their students, including their online students. The School of Nuclear Science and Engineering (NSE) at Oregon State University, in partnership with SpectralLabs company, has developed an Advanced Virtual Radiation Detection and Measurement Lab (AVR-DML) to offer a fully online Master of Health Physics program. It was essential for us to use a system that could simulate nuclear modules that accurately replicate the underlying physics, the nature of radiation and radiation transport, and the mechanics of the instrumentations used in the real radiation detection lab. It was all accomplished using a Realistic, Adaptive, Interactive Learning System (RAILS). RAILS is a comprehensive software simulation-based learning system for use in training. It is comprised of a web-based learning management system that is located on a central server, as well as a 3D-simulation package that is downloaded locally to user machines. Users will find that the graphics, animations, and sounds in RAILS create a realistic, immersive environment to practice detecting different radiation sources. These features allow students to coexist, interact and engage with a real STEM lab in all its dimensions. It enables them to feel like they are in a real lab environment and to see the same system they would in a lab. Unique interactive interfaces were designed and developed by integrating all the tools and equipment needed to run each lab. These interfaces provide students full functionality for data collection, changing the experimental setup, and live data collection with real-time updates for each experiment. Students can manually do all experimental setups and parameter changes in this lab. Experimental results can then be tracked and analyzed in an oscilloscope, a multi-channel analyzer, or a single-channel analyzer (SCA). The advanced virtual radiation detection and measurement laboratory developed in this study enabled the NSE school to offer a fully online MHP program. This flexibility of course modality helped us to attract more non-traditional students, including international students. It is a valuable educational tool as students can walk around the virtual lab, make mistakes, and learn from them. They have an unlimited amount of time to repeat and engage in experiments. This lab will also help us speed up training in nuclear science and engineering.

Keywords: advanced radiation detection and measurement, virtual laboratory, realistic adaptive interactive learning system (rails), online education in stem fields, student engagement, stem online education, stem laboratory, online engineering education

Procedia PDF Downloads 77
5708 Glaucoma Detection in Retinal Tomography Using the Vision Transformer

Authors: Sushish Baral, Pratibha Joshi, Yaman Maharjan

Abstract:

Glaucoma is a chronic eye condition that causes vision loss that is irreversible. Early detection and treatment are critical to prevent vision loss because it can be asymptomatic. For the identification of glaucoma, multiple deep learning algorithms are used. Transformer-based architectures, which use the self-attention mechanism to encode long-range dependencies and acquire extremely expressive representations, have recently become popular. Convolutional architectures, on the other hand, lack knowledge of long-range dependencies in the image due to their intrinsic inductive biases. The aforementioned statements inspire this thesis to look at transformer-based solutions and investigate the viability of adopting transformer-based network designs for glaucoma detection. Using retinal fundus images of the optic nerve head to develop a viable algorithm to assess the severity of glaucoma necessitates a large number of well-curated images. Initially, data is generated by augmenting ocular pictures. After that, the ocular images are pre-processed to make them ready for further processing. The system is trained using pre-processed images, and it classifies the input images as normal or glaucoma based on the features retrieved during training. The Vision Transformer (ViT) architecture is well suited to this situation, as it allows the self-attention mechanism to utilise structural modeling. Extensive experiments are run on the common dataset, and the results are thoroughly validated and visualized.

Keywords: glaucoma, vision transformer, convolutional architectures, retinal fundus images, self-attention, deep learning

Procedia PDF Downloads 181
5707 Optimal Pricing Based on Real Estate Demand Data

Authors: Vanessa Kummer, Maik Meusel

Abstract:

Real estate demand estimates are typically derived from transaction data. However, in regions with excess demand, transactions are driven by supply and therefore do not indicate what people are actually looking for. To estimate the demand for housing in Switzerland, search subscriptions from all important Swiss real estate platforms are used. These data do, however, suffer from missing information—for example, many users do not specify how many rooms they would like or what price they would be willing to pay. In economic analyses, it is often the case that only complete data is used. Usually, however, the proportion of complete data is rather small which leads to most information being neglected. Also, the data might have a strong distortion if it is complete. In addition, the reason that data is missing might itself also contain information, which is however ignored with that approach. An interesting issue is, therefore, if for economic analyses such as the one at hand, there is an added value by using the whole data set with the imputed missing values compared to using the usually small percentage of complete data (baseline). Also, it is interesting to see how different algorithms affect that result. The imputation of the missing data is done using unsupervised learning. Out of the numerous unsupervised learning approaches, the most common ones, such as clustering, principal component analysis, or neural networks techniques are applied. By training the model iteratively on the imputed data and, thereby, including the information of all data into the model, the distortion of the first training set—the complete data—vanishes. In a next step, the performances of the algorithms are measured. This is done by randomly creating missing values in subsets of the data, estimating those values with the relevant algorithms and several parameter combinations, and comparing the estimates to the actual data. After having found the optimal parameter set for each algorithm, the missing values are being imputed. Using the resulting data sets, the next step is to estimate the willingness to pay for real estate. This is done by fitting price distributions for real estate properties with certain characteristics, such as the region or the number of rooms. Based on these distributions, survival functions are computed to obtain the functional relationship between characteristics and selling probabilities. Comparing the survival functions shows that estimates which are based on imputed data sets do not differ significantly from each other; however, the demand estimate that is derived from the baseline data does. This indicates that the baseline data set does not include all available information and is therefore not representative for the entire sample. Also, demand estimates derived from the whole data set are much more accurate than the baseline estimation. Thus, in order to obtain optimal results, it is important to make use of all available data, even though it involves additional procedures such as data imputation.

Keywords: demand estimate, missing-data imputation, real estate, unsupervised learning

Procedia PDF Downloads 273
5706 Cultural Dynamics in Online Consumer Behavior: Exploring Cross-Country Variances in Review Influence

Authors: Eunjung Lee

Abstract:

This research investigates the intricate connection between cultural differences and online consumer behaviors by integrating Hofstede's Cultural Dimensions theory with analysis methodologies such as text mining, data mining, and topic analysis. Our aim is to provide a comprehensive understanding of how national cultural differences influence individuals' behaviors when engaging with online reviews. To ensure the relevance of our investigation, we systematically analyze and interpret the cultural nuances influencing online consumer behaviors, especially in the context of online reviews. By anchoring our research in Hofstede's Cultural Dimensions theory, we seek to offer valuable insights for marketers to tailor their strategies based on the cultural preferences of diverse global consumer bases. In our methodology, we employ advanced text mining techniques to extract insights from a diverse range of online reviews gathered globally for a specific product or service like Netflix. This approach allows us to reveal hidden cultural cues in the language used by consumers from various backgrounds. Complementing text mining, data mining techniques are applied to extract meaningful patterns from online review datasets collected from different countries, aiming to unveil underlying structures and gain a deeper understanding of the impact of cultural differences on online consumer behaviors. The study also integrates topic analysis to identify recurring subjects, sentiments, and opinions within online reviews. Marketers can leverage these insights to inform the development of culturally sensitive strategies, enhance target audience segmentation, and refine messaging approaches aligned with cultural preferences. Anchored in Hofstede's Cultural Dimensions theory, our research employs sophisticated methodologies to delve into the intricate relationship between cultural differences and online consumer behaviors. Applied to specific cultural dimensions, such as individualism vs. collectivism, masculinity vs. femininity, uncertainty avoidance, and long-term vs. short-term orientation, the study uncovers nuanced insights. For example, in exploring individualism vs. collectivism, we examine how reviewers from individualistic cultures prioritize personal experiences while those from collectivistic cultures emphasize communal opinions. Similarly, within masculinity vs. femininity, we investigate whether distinct topics align with cultural notions, such as robust features in masculine cultures and user-friendliness in feminine cultures. Examining information-seeking behaviors under uncertainty avoidance reveals how cultures differ in seeking detailed information or providing succinct reviews based on their comfort with ambiguity. Additionally, in assessing long-term vs. short-term orientation, the research explores how cultural focus on enduring benefits or immediate gratification influences reviews. These concrete examples contribute to the theoretical enhancement of Hofstede's Cultural Dimensions theory, providing a detailed understanding of cultural impacts on online consumer behaviors. As online reviews become increasingly crucial in decision-making, this research not only contributes to the academic understanding of cultural influences but also proposes practical recommendations for enhancing online review systems. Marketers can leverage these findings to design targeted and culturally relevant strategies, ultimately enhancing their global marketing effectiveness and optimizing online review systems for maximum impact.

Keywords: comparative analysis, cultural dimensions, marketing intelligence, national culture, online consumer behavior, text mining

Procedia PDF Downloads 33
5705 Impact of Integrated Signals for Doing Human Activity Recognition Using Deep Learning Models

Authors: Milagros Jaén-Vargas, Javier García Martínez, Karla Miriam Reyes Leiva, María Fernanda Trujillo-Guerrero, Francisco Fernandes, Sérgio Barroso Gonçalves, Miguel Tavares Silva, Daniel Simões Lopes, José Javier Serrano Olmedo

Abstract:

Human Activity Recognition (HAR) is having a growing impact in creating new applications and is responsible for emerging new technologies. Also, the use of wearable sensors is an important key to exploring the human body's behavior when performing activities. Hence, the use of these dispositive is less invasive and the person is more comfortable. In this study, a database that includes three activities is used. The activities were acquired from inertial measurement unit sensors (IMU) and motion capture systems (MOCAP). The main objective is differentiating the performance from four Deep Learning (DL) models: Deep Neural Network (DNN), Convolutional Neural Network (CNN), Recurrent Neural Network (RNN) and hybrid model Convolutional Neural Network-Long Short-Term Memory (CNN-LSTM), when considering acceleration, velocity and position and evaluate if integrating the IMU acceleration to obtain velocity and position represent an increment in performance when it works as input to the DL models. Moreover, compared with the same type of data provided by the MOCAP system. Despite the acceleration data is cleaned when integrating, results show a minimal increase in accuracy for the integrated signals.

Keywords: HAR, IMU, MOCAP, acceleration, velocity, position, feature maps

Procedia PDF Downloads 85
5704 The Intersection of Disability, Race and Gender in Keah Brown's 'The Pretty One'

Authors: Mehena Fedoul

Abstract:

This paper examines the intersection of race, gender, and disability through a Critical disability race theory and black feminist disability perspective in Keah Brown's memoir, "The Pretty One." The background of the study highlights the significance of intersectionality in understanding the multifaceted experiences of individuals who navigate multiple marginalized identities. The study contributes to the underrepresented field of disability studies from a Critical race and black feminist perspectives, shedding light on the unique challenges and resilience of black disabled women. The study employs a qualitative analysis of Keah Brown's memoir as a primary text. Drawing on intersectionality theory and black feminist disability scholarship, the analysis focuses on how Brown's memoir illuminates the ways in which her race, gender, and disability intersect and shape her lived experiences. The analysis reveals how Brown's memoir challenges traditional notions of disability, beauty, and empowerment through her unapologetic celebration of her blackness, femaleness, and disability. The major findings of the study indicate that Brown's memoir provides a powerful narrative of the complexity, uniqueness and richness of the lived experiences of black disabled women. It demonstrates how the intersectionality of race, gender, and disability shapes Brown's identity, body image, relationships, and societal interactions. The paper also highlights how Brown's memoir emphasizes the importance of inclusivity and intersectionality in understanding and addressing the challenges faced by black disabled women. In conclusion, this study offers a critical analysis of the intersection of race, gender, and disability in Keah Brown's memoir, "The Pretty One," from a black feminist disability perspective. It contributes to the growing body of literature that recognizes the significance of intersectionality in understanding the experiences of marginalized individuals in the disability community. The study underscores the need for more inclusive and intersectional perspectives in disability studies and advocates for greater recognition of the voices and experiences of black disabled women in academic and societal discourse.

Keywords: Intersectionality, black feminism, disability studies, keah brown

Procedia PDF Downloads 63
5703 Landfill Site Selection Using Multi-Criteria Decision Analysis A Case Study for Gulshan-e-Iqbal Town, Karachi

Authors: Javeria Arain, Saad Malik

Abstract:

The management of solid waste is a crucial and essential aspect of urban environmental management especially in a city with an ever increasing population such as Karachi. The total amount of municipal solid waste generated from Gulshan e Iqbal town on average is 444.48 tons per day and landfill sites are a widely accepted solution for final disposal of this waste. However, an improperly selected site can have immense environmental, economical and ecological impacts. To select an appropriate landfill site a number of factors should be kept into consideration to minimize the potential hazards of solid waste. The purpose of this research is to analyse the study area for the construction of an appropriate landfill site for disposal of municipal solid waste generated from Gulshan e-Iqbal Town by using geospatial techniques considering hydrological, geological, social and geomorphological factors. This was achieved using analytical hierarchy process and fuzzy analysis as a decision support tool with integration of geographic information sciences techniques. Eight most critical parameters, relevant to the study area, were selected. After generation of thematic layers for each parameter, overlay analysis was performed in ArcGIS 10.0 software. The results produced by both methods were then compared with each other and the final suitability map using AHP shows that 19% of the total area is Least Suitable, 6% is Suitable but avoided, 46% is Moderately Suitable, 26% is Suitable, 2% is Most Suitable and 1% is Restricted. In comparison the output map of fuzzy set theory is not in crisp logic rather it provides an output map with a range of 0-1, where 0 indicates least suitable and 1 indicates most suitable site. Considering the results it is deduced that the northern part of the city is appropriate for constructing the landfill site though a final decision for an optimal site could be made after field survey and considering economical and political factors.

Keywords: Analytical Hierarchy Process (AHP), fuzzy set theory, Geographic Information Sciences (GIS), Multi-Criteria Decision Analysis (MCDA)

Procedia PDF Downloads 490
5702 Health Trajectory Clustering Using Deep Belief Networks

Authors: Farshid Hajati, Federico Girosi, Shima Ghassempour

Abstract:

We present a Deep Belief Network (DBN) method for clustering health trajectories. Deep Belief Network (DBN) is a deep architecture that consists of a stack of Restricted Boltzmann Machines (RBM). In a deep architecture, each layer learns more complex features than the past layers. The proposed method depends on DBN in clustering without using back propagation learning algorithm. The proposed DBN has a better a performance compared to the deep neural network due the initialization of the connecting weights. We use Contrastive Divergence (CD) method for training the RBMs which increases the performance of the network. The performance of the proposed method is evaluated extensively on the Health and Retirement Study (HRS) database. The University of Michigan Health and Retirement Study (HRS) is a nationally representative longitudinal study that has surveyed more than 27,000 elderly and near-elderly Americans since its inception in 1992. Participants are interviewed every two years and they collect data on physical and mental health, insurance coverage, financial status, family support systems, labor market status, and retirement planning. The dataset is publicly available and we use the RAND HRS version L, which is easy to use and cleaned up version of the data. The size of sample data set is 268 and the length of the trajectories is equal to 10. The trajectories do not stop when the patient dies and represent 10 different interviews of live patients. Compared to the state-of-the-art benchmarks, the experimental results show the effectiveness and superiority of the proposed method in clustering health trajectories.

Keywords: health trajectory, clustering, deep learning, DBN

Procedia PDF Downloads 355
5701 Nursing Education in the Pandemic Time: Case Study

Authors: Jaana Sepp, Ulvi Kõrgemaa, Kristi Puusepp, Õie Tähtla

Abstract:

COVID-19 was officially recognized as a pandemic in late 2019 by the WHO, and it has led to changes in the education sector. Educational institutions were closed, and most schools adopted distance learning. Estonia is known as a digitally well-developed country. Based on that, in the pandemic time, nursing education continued, and new technological solutions were implemented. To provide nursing education, special focus was paid on quality and flexibility. The aim of this paper is to present administrative, digital, and technological solutions which support Estonian nursing educators to continue the study process in the pandemic time and to develop a sustainable solution for nursing education for the future. This paper includes the authors’ analysis of the documents and decisions implemented in the institutions through the pandemic time. It is a case study of Estonian nursing educators. Results of the analysis show that the implementation of distance learning principles challenges the development of innovative strategies and technics for the assessment of student performance and educational outcomes and implement new strategies to encourage student engagement in the virtual classroom. Additionally, hospital internships were canceled, and the simulation approach was deeply implemented as a new opportunity to develop and assess students’ practical skills. There are many other technical and administrative changes that have also been carried out, such as students’ support and assessment systems, the designing and conducting of hybrid and blended studies, etc. All services were redesigned and made more available, individual, and flexible. Hence, the feedback system was changed, the information was collected in parallel with educational activities. Experiences of nursing education during the pandemic time are widely presented in scientific literature. However, to conclude our study, authors have found evidence that solutions implemented in Estonian nursing education allowed the students to graduate within the nominal study period without any decline in education quality. Operative information system and flexibility provided the minimum distance between the students, support, and academic staff, and likewise, the changes were implemented quickly and efficiently. Institution memberships were updated with the appropriate information, and it positively affected their satisfaction, motivation, and commitment. We recommend that the feedback process and the system should be permanently changed in the future to place all members in the same information area, redefine the hospital internship process, implement hybrid learning, as well as to improve the communication system between stakeholders inside and outside the organization. The main limitation of this study relates to the size of Estonia. Nursing education is provided by two institutions only, and similarly, the number of students is low. The result could be generated to the institutions with a similar size and administrative system. In the future, the relationship between nurses’ performance and organizational outcomes should be deeply investigated and influences of the pandemic time education analyzed at workplaces.

Keywords: hybrid learning, nursing education, nursing, COVID-19

Procedia PDF Downloads 114
5700 Challenges in E-Government: Conceptual Views and Solutions

Authors: Rasim Alguliev, Farhad Yusifov

Abstract:

Considering the international experience, conceptual and architectural principles of forming of electron government are researched and some suggestions were made. The assessment of monitoring of forming processes of electron government, intellectual analysis of web-resources, provision of information security, electron democracy problems were researched, conceptual approaches were suggested. By taking into consideration main principles of electron government theory, important research directions were specified.

Keywords: electron government, public administration, information security, web-analytics, social networks, data mining

Procedia PDF Downloads 454
5699 A Kierkegaardian Reading of Iqbal's Poetry as a Communicative Act

Authors: Sevcan Ozturk

Abstract:

The overall aim of this paper is to present a Kierkegaardian approach to Iqbal’s use of literature as a form of communication. Despite belonging to different historical, cultural, and religious backgrounds, the philosophical approaches of Soren Kierkegaard, ‘the father of existentialism,' and Muhammad Iqbal ‘the spiritual father of Pakistan’ present certain parallels. Both Kierkegaard and Iqbal take human existence as the starting point for their reflections, emphasise the subject of becoming genuine religious personalities, and develop a notion of the self. While doing these they both adopt parallel methods, employ literary techniques and poetical forms, and use their literary works as a form of communication. The problem is that Iqbal does not provide a clear account of his method as Kierkegaard does in his works. As a result, Iqbal’s literary approach appears to be a collection of contradictions. This is mainly because despite he writes most of his works in the poetical form, he condemns all kinds of art including poetry. Moreover, while attacking on Islamic mysticism, he, at the same time, uses classical literary forms, and a number of traditional mystical, poetic symbols. This paper will argue that the contradictions found in Iqbal’s approach are actually a significant part of Iqbal’s way of communicating his reader. It is the contention of this paper that with the help of the parallels between the literary and philosophical theories of Kierkegaard and Iqbal, the application of Kierkegaard’s method to Iqbal’s use of poetry as a communicative act will make it possible to dispel the seeming ambiguities in Iqbal’s literary approach. The application of Kierkegaard’s theory to Iqbal’s literary method will include an analysis of the main principles of Kierkegaard’s own literary technique of ‘indirect communication,' which is a crucial term of his existentialist philosophy. Second, the clash between what Iqbal’s says about art and poetry and what he does will be highlighted in the light of Kierkegaardian theory of indirect communication. It will be argued that Iqbal’s literary technique can be considered as a form of ‘indirect communication,' and that reading his technique in this way helps on dispelling the contradictions in his approach. It is hoped that this paper will cultivate a dialogue between those who work in the fields of comparative philosophy Kierkegaard studies, existentialism, contemporary Islamic thought, Iqbal studies, and literary criticism.

Keywords: comparative philosophy, existentialism, indirect communication, intercultural philosophy, literary communication, Muhammad Iqbal, Soren Kierkegaard

Procedia PDF Downloads 314
5698 DenseNet and Autoencoder Architecture for COVID-19 Chest X-Ray Image Classification and Improved U-Net Lung X-Ray Segmentation

Authors: Jonathan Gong

Abstract:

Purpose AI-driven solutions are at the forefront of many pathology and medical imaging methods. Using algorithms designed to better the experience of medical professionals within their respective fields, the efficiency and accuracy of diagnosis can improve. In particular, X-rays are a fast and relatively inexpensive test that can diagnose diseases. In recent years, X-rays have not been widely used to detect and diagnose COVID-19. The under use of Xrays is mainly due to the low diagnostic accuracy and confounding with pneumonia, another respiratory disease. However, research in this field has expressed a possibility that artificial neural networks can successfully diagnose COVID-19 with high accuracy. Models and Data The dataset used is the COVID-19 Radiography Database. This dataset includes images and masks of chest X-rays under the labels of COVID-19, normal, and pneumonia. The classification model developed uses an autoencoder and a pre-trained convolutional neural network (DenseNet201) to provide transfer learning to the model. The model then uses a deep neural network to finalize the feature extraction and predict the diagnosis for the input image. This model was trained on 4035 images and validated on 807 separate images from the ones used for training. The images used to train the classification model include an important feature: the pictures are cropped beforehand to eliminate distractions when training the model. The image segmentation model uses an improved U-Net architecture. This model is used to extract the lung mask from the chest X-ray image. The model is trained on 8577 images and validated on a validation split of 20%. These models are calculated using the external dataset for validation. The models’ accuracy, precision, recall, f1-score, IOU, and loss are calculated. Results The classification model achieved an accuracy of 97.65% and a loss of 0.1234 when differentiating COVID19-infected, pneumonia-infected, and normal lung X-rays. The segmentation model achieved an accuracy of 97.31% and an IOU of 0.928. Conclusion The models proposed can detect COVID-19, pneumonia, and normal lungs with high accuracy and derive the lung mask from a chest X-ray with similarly high accuracy. The hope is for these models to elevate the experience of medical professionals and provide insight into the future of the methods used.

Keywords: artificial intelligence, convolutional neural networks, deep learning, image processing, machine learning

Procedia PDF Downloads 116
5697 The Social Justice of Movement: Undocumented Immigrant Coalitions in the United States

Authors: Libia Jiménez Chávez

Abstract:

This is a study of freedom riders and their courageous journey for civil rights, but the year was not 1961. It was 2003. This paper chronicles the emergence of a new civil rights movement for immigrant rights through an oral history of the 2003 U.S. Immigrant Workers Freedom Ride (IWFR). During the height of the post-9/11 immigrant repression, a bloc of organizations inspired by the Civil Rights Movement of the 1960s mobilized 900 multinational immigrants and their allies in the fight for legal status, labor protections, family reunification, and civil rights. The activists visited over 100 U.S. cities, met with Congressional leaders in the nation’s capital, and led a rally of over 50,000 people in New York City. This unified effort set the groundwork for the national May Day immigration protests of 2006. Movements can be characterized in two distinct ways: physical movement and social movements. In the past, historians have considered immigrants both as people and as participants in social movements. In contrast, studies of recent migrants tend to say little about their involvement in immigrant political mobilizations. The dominant literature on immigration portrays immigrants as objects of exclusion, border enforcement, detention, and deportation instead of strategic political actors. This paper aims to change this perception. It considers the Freedom Riders both as immigrants who were literally on the move and as participants in a social movement. Through interviews with participants and archival video footage housed at the University of California Los Angeles, it is possible to study this mobile protest as a movement. This contemporary immigrant struggle is an opportunity to explore the makeup and development of a heterogenous immigrant coalition and consider the relationship between population movements and social justice. In addition to oral histories and archival research, the study will utilize social movement literature, U.S. immigration and labor history, and Undocumented Critical Theory to expand the historiography of immigrant social movements in America.

Keywords: civil rights, immigrant social movements, undocumented communities, undocumented critical theory

Procedia PDF Downloads 147
5696 Research and Development of Methodology, Tools, Techniques and Methods to Analyze and Design Interface, Media, Pedagogy for Educational Topics to be Delivered via Mobile Technology

Authors: Shimaa Nagro, Russell Campion

Abstract:

Mobile devices are becoming ever more widely available, with growing functionality, and they are increasingly used as enabling technology to give students access to educational material anytime and anywhere. However, the design of educational material's user interfaces for mobile devices is beset by many unresolved research problems such as those arising from constraints associated with mobile devices or from issues linked to effective learning. The proposed research aims to produce: (i) a method framework for the design and evaluation of educational material’s interfaces to be delivered on mobile devices, in multimedia form based on Human Computer Interaction strategies; and (ii) a software tool implemented as a fast-track alternative to use the method framework in full. The investigation will combine qualitative and quantitative methods, including interviews and questionnaires for data collection and three case studies for validating the method framework. The method framework is a framework to enable an educational designer to effectively and efficiently create educational multimedia interfaces to be used on mobile devices by following a particular methodology that contains practical and usable tools and techniques. It is a method framework that accepts any educational material in its final lesson plan and deals with this plan as a static element, it will not suggest any changes in any information given in the lesson plan but it will help the instructor to design his final lesson plan in a multimedia format to be presented in mobile devices.

Keywords: mobile learning, M-Learn, HCI, educational multimedia, interface design

Procedia PDF Downloads 359
5695 Desk Graffiti as Art, Archive or Collective Knowledge Sharing: A Case Study of Schools in Addis Ababa, Ethiopia

Authors: Behailu Bezabih Ayele

Abstract:

Illustrative expressions in art education and in overall learning are being given increasing attention in the transmission of knowledge. The objective of this paper, therefore, is to present an analysis of graffiti on school desks-a way of smuggling knowledge on the edge of classroom education and learning. The methodological approach focuses on the systematic collection and selection of desk graffiti. Four schools are chosen to reflect socioeconomic status and gender composition. The analysis focused on the categorization of graffiti by genre. This was followed by an analysis of the style, intensity as well as content of the messages in terms of overall social impacts. The paper grounds the analysis by reviewing the literature on modern education and art education in the Ethiopian context, as well as the place of desk graffiti. The findings generally show that the school desks and the school environment, by and large, have managed to serve as vessels through which formal and informal knowledge is acquired, transmitted, engrained into the students and transformed into messages by the students. The desks have also apparently served as a springboard to maximize the interfaces between several ideas and disciplines and communications. However, the very fact that the desks serve as massive channels of expression and knowledge transmission also points to a lack of breadth availability of channels of expression, perhaps confounding the ability of classrooms as means of outlet of expression and documentation for the students. This points to the need for efforts in education policy and funding of artistic endeavors for young students.

Keywords: artistic expression, desk graffiti, education, school children, Ethiopia

Procedia PDF Downloads 57
5694 Deep Learning-Based Object Detection on Low Quality Images: A Case Study of Real-Time Traffic Monitoring

Authors: Jean-Francois Rajotte, Martin Sotir, Frank Gouineau

Abstract:

The installation and management of traffic monitoring devices can be costly from both a financial and resource point of view. It is therefore important to take advantage of in-place infrastructures to extract the most information. Here we show how low-quality urban road traffic images from cameras already available in many cities (such as Montreal, Vancouver, and Toronto) can be used to estimate traffic flow. To this end, we use a pre-trained neural network, developed for object detection, to count vehicles within images. We then compare the results with human annotations gathered through crowdsourcing campaigns. We use this comparison to assess performance and calibrate the neural network annotations. As a use case, we consider six months of continuous monitoring over hundreds of cameras installed in the city of Montreal. We compare the results with city-provided manual traffic counting performed in similar conditions at the same location. The good performance of our system allows us to consider applications which can monitor the traffic conditions in near real-time, making the counting usable for traffic-related services. Furthermore, the resulting annotations pave the way for building a historical vehicle counting dataset to be used for analysing the impact of road traffic on many city-related issues, such as urban planning, security, and pollution.

Keywords: traffic monitoring, deep learning, image annotation, vehicles, roads, artificial intelligence, real-time systems

Procedia PDF Downloads 182
5693 Innovations and Challenges: Multimodal Learning in Cybersecurity

Authors: Tarek Saadawi, Rosario Gennaro, Jonathan Akeley

Abstract:

There is rapidly growing demand for professionals to fill positions in Cybersecurity. This is recognized as a national priority both by government agencies and the private sector. Cybersecurity is a very wide technical area which encompasses all measures that can be taken in an electronic system to prevent criminal or unauthorized use of data and resources. This requires defending computers, servers, networks, and their users from any kind of malicious attacks. The need to address this challenge has been recognized globally but is particularly acute in the New York metropolitan area, home to some of the largest financial institutions in the world, which are prime targets of cyberattacks. In New York State alone, there are currently around 57,000 jobs in the Cybersecurity industry, with more than 23,000 unfilled positions. The Cybersecurity Program at City College is a collaboration between the Departments of Computer Science and Electrical Engineering. In Fall 2020, The City College of New York matriculated its first students in theCybersecurity Master of Science program. The program was designed to fill gaps in the previous offerings and evolved out ofan established partnership with Facebook on Cybersecurity Education. City College has designed a program where courses, curricula, syllabi, materials, labs, etc., are developed in cooperation and coordination with industry whenever possible, ensuring that students graduating from the program will have the necessary background to seamlessly segue into industry jobs. The Cybersecurity Program has created multiple pathways for prospective students to obtain the necessary prerequisites to apply in order to build a more diverse student population. The program can also be pursued on a part-time basis which makes it available to working professionals. Since City College’s Cybersecurity M.S. program was established to equip students with the advanced technical skills needed to thrive in a high-demand, rapidly-evolving field, it incorporates a range of pedagogical formats. From its outset, the Cybersecurity program has sought to provide both the theoretical foundations necessary for meaningful work in the field along with labs and applied learning projects aligned with skillsets required by industry. The efforts have involved collaboration with outside organizations and with visiting professors designing new courses on topics such as Adversarial AI, Data Privacy, Secure Cloud Computing, and blockchain. Although the program was initially designed with a single asynchronous course in the curriculum with the rest of the classes designed to be offered in-person, the advent of the COVID-19 pandemic necessitated a move to fullyonline learning. The shift to online learning has provided lessons for future development by providing examples of some inherent advantages to the medium in addition to its drawbacks. This talk will address the structure of the newly-implemented Cybersecurity Master’s Program and discuss the innovations, challenges, and possible future directions.

Keywords: cybersecurity, new york, city college, graduate degree, master of science

Procedia PDF Downloads 128
5692 Effect of Formative Evaluation with Feedback on Students Economics Achievement in Secondary Education

Authors: Salihu Abdullahi Galle

Abstract:

Students' performance in Economics in schools and on standardized exams in Nigeria has been worrying throughout the years, owing to some teachers' use of conventional and lecture teaching methods. Other obstacles include a lack of training, standardized testing pressure, and aversion to change, all of which can have an impact on students' cognitive ability in Economics and future careers. The researchers employed formative evaluation with feedback (FEFB) to support the teaching and learning process by providing constant feedback to both teachers and students. The researchers employed a quasi-experimental research design to examine two teaching methods (FEFB and traditional). The pre-test and post-test interaction effects were evaluated between students in the experimental group (FEFB) and those in the conventional group. The interaction effects of pre-test and post-test on male and female in the two groups were also examined, with 90 participants. The findings show that students exposed to a FEFB-based teaching approach outperform pupils taught in a traditional classroom setting, and there is no gender interaction effect between the two groups. In light of these findings, the researchers urge that Economics teachers employ FEFB during teaching and learning to ensure timely feedback, and that policymakers ensure that Economics teachers receive training and re-training on FEFB approaches.

Keywords: formative evaluation with feedback (FEFB), students, economics achievement, secondary education

Procedia PDF Downloads 36
5691 Interdisciplinarity as a Regular Pedagogical Practice in the Classrooms

Authors: Catarina Maria Neto Da Cruz, Ana Maria Reis D’Azevedo Breda

Abstract:

The world is changing and, consequently, the young people need more sophisticated tools and skills to lead with the world’s complexity. The Organisation for Economic Co-operation and Development Learning Framework 2030 suggests an interdisciplinary knowledge as a principle for the future of education systems. In the curricular document Portuguese about the profile of students leaving compulsory education, the critical thinking and creative thinking are pointed out as skills to be developed, which imply the interconnection of different knowledge, applying it in different contexts and learning areas. Unlike primary school teachers, teachers specialized in a specific area lead to more difficulties in the implementation of interdisciplinary approaches in the classrooms and, despite the effort, the interdisciplinarity is not a common practice in schools. Statement like "Mathematics is everywhere" is unquestionable, however, many math teachers show difficulties in presenting such evidence in their classes. Mathematical modelling and problems in real contexts are promising in the development of interdisciplinary pedagogical practices and in Portugal there is a continuous training offer to contribute to the development of teachers in terms of their pedagogical approaches. But when teachers find themselves in the classroom, without a support, do they feel able to implement interdisciplinary practices? In this communication we will try to approach this issue through a case study involving a group of Mathematics teachers, who attended a training aimed at stimulating interdisciplinary practices in real contexts, namely related to the COVID-19 pandemic.

Keywords: education, mathematics, teacher training, interdisciplinarity

Procedia PDF Downloads 75
5690 Estimation of Effective Mechanical Properties of Linear Elastic Materials with Voids Due to Volume and Surface Defects

Authors: Sergey A. Lurie, Yury O. Solyaev, Dmitry B. Volkov-Bogorodsky, Alexander V. Volkov

Abstract:

The media with voids is considered and the method of the analytical estimation of the effective mechanical properties in the theory of elastic materials with voids is proposed. The variational model of the porous media is discussed, which is based on the model of the media with fields of conserved dislocations. It is shown that this model is fully consistent with the known model of the linear elastic materials with voids. In the present work, the generalized model of the porous media is proposed in which the specific surface properties are associated with the field of defects-pores in the volume of the deformed body. Unlike typical surface elasticity model, the strain energy density of the considered model includes the special part of the surface energy with the quadratic form of the free distortion tensor. In the result, the non-classical boundary conditions take modified form of the balance equations of volume and surface stresses. The analytical approach is proposed in the present work which allows to receive the simple enough engineering estimations for effective characteristics of the media with free dilatation. In particular, the effective flexural modulus and Poisson's ratio are determined for the problem of a beam pure bending. Here, the known voids elasticity solution was expanded on the generalized model with the surface effects. Received results allow us to compare the deformed state of the porous beam with the equivalent classic beam to introduce effective bending rigidity. Obtained analytical expressions for the effective properties depend on the thickness of the beam as a parameter. It is shown that the flexural modulus of the porous beam is decreased with an increasing of its thickness and the effective Poisson's ratio of the porous beams can take negative values for the certain values of the model parameters. On the other hand, the effective shear modulus is constant under variation of all values of the non-classical model parameters. Solutions received for a beam pure bending and the hydrostatic loading of the porous media are compared. It is shown that an analytical estimation for the bulk modulus of the porous material under hydrostatic compression gives an asymptotic value for the effective bulk modulus of the porous beam in the case of beam thickness increasing. Additionally, it is shown that the scale effects appear due to the surface properties of the porous media. Obtained results allow us to offer the procedure of an experimental identification of the non-classical parameters in the theory of the linear elastic materials with voids based on the bending tests for samples with different thickness. Finally, the problem of implementation of the Saint-Venant hypothesis for the transverse stresses in the porous beam are discussed. These stresses are different from zero in the solution of the voids elasticity theory, but satisfy the integral equilibrium equations. In this work, the exact value of the introduced surface parameter was found, which provides the vanishing of the transverse stresses on the free surfaces of a beam.

Keywords: effective properties, scale effects, surface defects, voids elasticity

Procedia PDF Downloads 393
5689 Analyzing the Street Pattern Characteristics on Young People’s Choice to Walk or Not: A Study Based on Accelerometer and Global Positioning Systems Data

Authors: Ebru Cubukcu, Gozde Eksioglu Cetintahra, Burcin Hepguzel Hatip, Mert Cubukcu

Abstract:

Obesity and overweight cause serious health problems. Public and private organizations aim to encourage walking in various ways in order to cope with the problem of obesity and overweight. This study aims to understand how the spatial characteristics of urban street pattern, connectivity and complexity influence young people’s choice to walk or not. 185 public university students in Izmir, the third largest city in Turkey, participated in the study. Each participant had worn an accelerometer and a global positioning (GPS) device for a week. The accelerometer device records data on the intensity of the participant’s activity at a specified time interval, and the GPS device on the activities’ locations. Combining the two datasets, activity maps are derived. These maps are then used to differentiate the participants’ walk trips and motor vehicle trips. Given that, the frequency of walk and motor vehicle trips are calculated at the street segment level, and the street segments are then categorized into two as ‘preferred by pedestrians’ and ‘preferred by motor vehicles’. Graph Theory-based accessibility indices are calculated to quantify the spatial characteristics of the streets in the sample. Six different indices are used: (I) edge density, (II) edge sinuosity, (III) eta index, (IV) node density, (V) order of a node, and (VI) beta index. T-tests show that the index values for the ‘preferred by pedestrians’ and ‘preferred by motor vehicles’ are significantly different. The findings indicate that the spatial characteristics of the street network have a measurable effect on young people’s choice to walk or not. Policy implications are discussed. This study is funded by the Scientific and Technological Research Council of Turkey, Project No: 116K358.

Keywords: graph theory, walkability, accessibility, street network

Procedia PDF Downloads 210
5688 Strengthening Adult Literacy Programs in Order to End Female Genital Mutilation to Achieve Sustainable Development Goals

Authors: Odenigbo Veronica Ngozi, Lorreta Chika Ukwuaba

Abstract:

This study focuses on how the strengthening adult literacy program can help accelerate transformative strategies to end female genital mutilation (FGM) in Nigeria, specifically in Nsukka Local Government Area. The research delves into the definition of FGM, adult literacy programs, and how to achieve ending FGM to attain Sustainable Development Goals (SDGs) in 2030. It further discusses the practice of FGM in Nigeria and emphasizes the statement of the problem. The main aim of the study is to investigate how strengthening adult literacy programs can help accelerate transformative strategies to end FGM in Nigeria and achieve SDGs in 2030. The researchers utilized a survey research design to conduct the study in Nsukka L.G.A. The population was composed of 26 facilitators and adult learners in five adult learning centers in the area. The entire population was used as a sample, and structured questionnaires were employed to elicit information. The items on the questionnaire were face-validated by three experts, and the reliability of the instrument was verified using Cronbach Alpha Reliability Technique. The research questions were analyzed using means and standard deviation while the hypothesis was tested at 0.05 level of degree of significance using a t-test. The findings show that through adult literacy program acceleration of transformative strategies, the practices of FGM can be ended. Strengthening adult literacy programs is a good channel to end or stop FGM through the knowledge and skill acquired from the learning centers. The theoretical importance of the study lies in the fact that it highlights the role of adult literacy programs in accelerating transformative strategies to combat harmful cultural practices such as FGM. It further supports the importance of education and knowledge in achieving sustainable development goals by 2030. Structured questionnaires were distributed to an entire population of 26 facilitators and adult learners in five adult learning centers in Nsukka L.G.A. The questionnaire items were face–validated by three experts, and the reliability of the instrument was verified using Cronbach Alpha Reliability Technique. The research questions were analyzed using means and standard deviation, while the hypothesis was tested using a t-test at 0.05 level of degree of significance. The study addressed the question of how strengthening adult literacy programs can help accelerate transformative strategies to end FGM in Nigeria and achieve SDGs by 2030. In conclusion, the study found that adult literacy is a good tool to end FGM in Nigeria. The recommendations were that government, non-governmental organizations (NGOs), Community-based organizations (CBOs), and individuals should support the funding and establishment of adult literacy centers in communities so as to reach every illiterate parent or individual and acquire the knowledge and skill needed to understand the negative effect of FGM in the life of a girl child.

Keywords: adult literacy, female genital mutilation, learning centers, SDGs, strengthening

Procedia PDF Downloads 56