Search results for: academic speed and accuracy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8920

Search results for: academic speed and accuracy

6490 Cracks Detection and Measurement Using VLP-16 LiDAR and Intel Depth Camera D435 in Real-Time

Authors: Xinwen Zhu, Xingguang Li, Sun Yi

Abstract:

Crack is one of the most common damages in buildings, bridges, roads and so on, which may pose safety hazards. However, cracks frequently happen in structures of various materials. Traditional methods of manual detection and measurement, which are known as subjective, time-consuming, and labor-intensive, are gradually unable to meet the needs of modern development. In addition, crack detection and measurement need be safe considering space limitations and danger. Intelligent crack detection has become necessary research. In this paper, an efficient method for crack detection and quantification using a 3D sensor, LiDAR, and depth camera is proposed. This method works even in a dark environment, which is usual in real-world applications. The LiDAR rapidly spins to scan the surrounding environment and discover cracks through lasers thousands of times per second, providing a rich, 3D point cloud in real-time. The LiDAR provides quite accurate depth information. The precision of the distance of each point can be determined within around  ±3 cm accuracy, and not only it is good for getting a precise distance, but it also allows us to see far of over 100m going with the top range models. But the accuracy is still large for some high precision structures of material. To make the depth of crack is much more accurate, the depth camera is in need. The cracks are scanned by the depth camera at the same time. Finally, all data from LiDAR and Depth cameras are analyzed, and the size of the cracks can be quantified successfully. The comparison shows that the minimum and mean absolute percentage error between measured and calculated width are about 2.22% and 6.27%, respectively. The experiments and results are presented in this paper.

Keywords: LiDAR, depth camera, real-time, detection and measurement

Procedia PDF Downloads 235
6489 Enhancing Embedded System Efficiency with Digital Signal Processing Cores

Authors: Anil Dhanawade, Akshay S., Harshal Lakesar

Abstract:

This paper presents a comprehensive analysis of the performance advantages offered by DSP (Digital Signal Processing) cores compared to traditional MCU (Microcontroller Unit) cores in the execution of various functions critical to real-time applications. The focus is on the integration of DSP functionalities, specifically in the context of motor control applications such as Field-Oriented Control (FOC), trigonometric calculations, back-EMF estimation, digital filtering, and high-resolution PWM generation. Through comparative analysis, it is demonstrated that DSP cores significantly enhance processing efficiency, achieving faster execution times for complex mathematical operations essential for precise torque and speed control. The study highlights the capabilities of DSP cores, including single-cycle Multiply-Accumulate (MAC) operations and optimized hardware for trigonometric functions, which collectively reduce latency and improve real-time performance. In contrast, MCU cores, while capable of performing similar tasks, typically exhibit longer execution times due to reliance on software-based solutions and lack of dedicated hardware acceleration. The findings underscore the critical role of DSP cores in applications requiring high-speed processing and low-latency response, making them indispensable in automotive, industrial, and robotics sectors. This work serves as a reference for future developments in embedded systems, emphasizing the importance of architecture choice in achieving optimal performance in demanding computational tasks.

Keywords: assembly code, DSP core, instruction set, MCU core

Procedia PDF Downloads 25
6488 Thermal and Hydraulic Design of Shell and Tube Heat Exchangers

Authors: Ahmed R. Ballil

Abstract:

Heat exchangers are devices used to transfer heat between two fluids. These devices are utilized in many engineering and industrial applications such as heating, cooling, condensation and boiling processes. The fluids might be in direct contact (mixed), or they separated by a solid wall to avoid mixing. In the present paper, interactive computer-aided design of shell and tube heat exchangers is developed using Visual Basic computer code as a framework. This design is based on the Bell-Delaware method, which is one of the very well known methods reported in the literature for the design of shell and tube heat exchangers. Physical properties for either the tube or the shell side fluids are internally evaluated by calling on an enormous data bank composed of more than a hundred fluid compounds. This contributes to increase the accuracy of the present design. The international system of units is considered in the developed computer program. The present design has an added feature of being capable of performing modification based upon a preset design criterion, such that an optimum design is obtained at satisfying constraints set either by the user or by the method itself. Also, the present code is capable of giving an estimate of the approximate cost of the heat exchanger based on the predicted surface area of the exchanger evaluated by the program. Finally, the present thermal and hydraulic design code is tested for accuracy and consistency against some of existed and approved designs of shell and tube heat exchangers.

Keywords: bell-delaware method, heat exchangers, shell and tube, thermal and hydraulic design

Procedia PDF Downloads 152
6487 The Library as a Metaphor: Perceptions, Evolution, and the Shifting Role in Society Through a Librarian's Lens

Authors: Nihar Kanta Patra, Akhtar Hussain

Abstract:

This comprehensive study, through the perspective of librarians, explores the library as a metaphor and its profound significance in representing knowledge and learning. It delves into how librarians perceive the library as a metaphor and the ways in which it symbolizes the acquisition, preservation, and dissemination of knowledge. The research investigates the most common metaphors used to describe libraries, as witnessed by librarians, and analyzes how these metaphors reflect the evolving role of libraries in society. Furthermore, the study examines how the library metaphor influences the perception of librarians regarding academic libraries as physical places and academic library websites as virtual spaces, exploring their potential for learning and exploration. It investigates the evolving nature of the library as a metaphor over time, as seen by librarians, considering the changing landscape of information and technology. The research explores the ways in which the library metaphor has expanded beyond its traditional representation, encompassing digital resources, online connectivity, and virtual realms, and provides insights into its potential evolution in the future. Drawing on the experiences of librarians in their interactions with library users, the study uncovers any specific cultural or generational differences in how people interpret or relate to the library as a metaphor. It sheds light on the diverse perspectives and interpretations of the metaphor based on cultural backgrounds, educational experiences, and technological familiarity. Lastly, the study investigates the evolving roles of libraries as observed by librarians and explores how these changing roles can influence the metaphors we use to represent them. It examines the dynamic nature of libraries as they adapt to societal needs, technological advancements, and new modes of information dissemination. By analyzing these various dimensions, this research provides a comprehensive understanding of the library as a metaphor through the lens of librarians, illuminating its significance, evolution, and its transformative impact on knowledge, learning, and the changing role of libraries in society.

Keywords: library, librarians, metaphor, perception

Procedia PDF Downloads 103
6486 Determination of Geogrid Reinforced Ballast Behavior Using Finite Element Modeling

Authors: Buğra Sinmez

Abstract:

In some countries, such as China, Turkey, andseveralEuropeanUnionnations, the therailwaypavementstructuralsystem has recently undergonerapid growth as a vital element of the transportation infrastructure, particularlyfortheuse of high-speed trains. It is vitaltoconsiderthe High-SpeedInfrastructureDemandwhendevelopingandconstructingtherailwaypavementstructure. HSRL can create more substantial ldifficultiestotheballastorbaselayer of regularlyusedballastedrailwaypavementsthanstandardrailwaytrains. The deterioration of the theballastorbaselayermayleadtosubstructuredegradation, which might lead to safety concerns and catastrophicincidents. As a result, the efficiency of railways will be impactedbylargecargoesorhigh-speed trains. A railwaypavement construction can be strengthened using geosyntheticmaterials in theballastorfoundationlayer as a countermeasure. However, there is still a need in the literature to quantifytheinfluence of geosynthetic materials, particularlygeogrid, on the mechanical responses of railwaypavementstructuresto HSRL loads which is essential knowledge in supporting the selection of appropriate material and geogridinstallationposition. As a result, the purpose of this research is to see how a geogridreinforcementlayermayaffectthekeyfeatures of a ballastedrailwaypavementstructure, with a particular focus on the materialtypeandgeogridplacementpositionthatmayassistreducethe rate of degradation of the therailwaypavementstructuresystem. Thisstudyusesnumericalmodeling in a genuinerailwaycontexttovalidatethebenefit of geogrid reinforcement. The usage of geogrids in the railway system has been thoroughly researched in the technical literature. Three distinct types of geogrid installed at two distinct positions (i.e.,withintheballastlayer, betweentheballastandthesub-ballast layer) within a railwaypavementconstructionwereevaluatedunder a variety of verticalwheelloadsusing a three-dimensional (3D) finite element model. As a result, fouralternativegeogridreinforcementsystemsweremodeledtoreflectdifferentconditions in the ballastedrailwaysystems (G0: no reinforcement; G1: reinforcedwithgeogridhavingthelowestdensityandYoung'smodulus; G2: reinforcedwithgeogridhavingtheintermediateYoung'smodulusanddensity; G3: reinforcedwithgeogridhavingthegreatestdensityandYoung'smodulus). Themechanicalreactions of the railway, such as verticalsurfacedeflection, maximumprimarystressandstrain, andmaximumshearstress, werestudiedandcomparedbetweenthefourgeogridreinforcementscenariosandfourverticalwheelloadlevels (i.e., 75, 100, 150, and 200 kN). Differences in the mechanical reactions of railwaypavementconstructionsowingtotheuse of differentgeogridmaterialsdemonstratethebenefits of suchgeosynthetics in ballast. In comparison to a non-reinforcedrailwaypavementconstruction, thereinforcedconstructionsfeaturedecreasedverticalsurfacedeflection, maximum shear stress at the sleeper-ballast contact, and maximum main stress at the bottom of the ballast layer. As a result, addinggeogridtotheballastlayerandbetweentheballastandsub-ballast layer in a ballastedrailwaypavementconstruction has beenfoundtolowercriticalshearand main stresses as well as verticalsurfacedeflection.

Keywords: geosynthetics, geogrid, railway, transportation

Procedia PDF Downloads 184
6485 Analytical Determination of Electromechanical Coupling Effects on Interlaminar Stresses of Generally Laminated Piezoelectric Plates

Authors: Atieh Andakhshideh, S. Maleki, Sayed Sadegh Marashi

Abstract:

In this paper, the interlaminar stresses of generally laminated piezoelectric plates are presented. The electromechanical coupling effect of the piezoelectric plate is considered and the governing equations and boundary conditions are derived using the principle of minimum total potential energy. The solution procedure is a three-dimensional multi-term extended Kantorovich method (3DMTEKM). The objective of this paper is to accurately study coupling influence on the edge effects of piezolaminated plates with finite dimensions, arbitrary lamination lay-ups and under uniform axial strain. These results can provide a benchmark for checking the accuracy of the other numerical method or two-dimensional laminate theories. To verify the accuracy of the 3DMTEKM, first examples are simplified to special cases such as cross-ply or symmetric laminations and are compared with other analytical solutions available in the literature. Excellent agreement is achieved in validation test and other numerical results are presented for general cases. Numerical examples indicate the singular behavior of interlaminar normal/shear stresses and electric field strength components near the edges of the piezolaminated plates. The coupling influence on the free edge effect with respect to lamination lay-ups of piezoelectric plate is studied in several examples.

Keywords: electromechanical coupling, generally laminated piezoelectric plates, Kantorovich method, edge effect, interlaminar stresses

Procedia PDF Downloads 151
6484 Evaluation of Clinical Decision Support System in Electronic Medical Record System: A Case of Malawi National Art Electronic Medical Record System

Authors: Pachawo Bisani, Goodall Nyirenda

Abstract:

The Malawi National Antiretroviral Therapy (NART) Electronic Medical Record (EMR) system was designed and developed with guidance from the Ministry of Health through the Department of HIV and AIDS (DHA) with the aim of supporting the management of HIV patient data and reporting in high prevalence ART clinics. As of 2021, the system has been scaled up to over 206 facilities across the country. The system is integrated with the clinical decision support system (CDSS) to assist healthcare providers in making a decision about an individual patient at a particular point in time. Despite NART EMR undergoing several evaluations and assessments, little has been done to evaluate the clinical decision support system in the NART EMR system. Hence, the study aimed to evaluate the use of CDSS in the NART EMR system in Malawi. The study adopted a mixed-method approach, and data was collected through interviews, observations, and questionnaires. The study has revealed that the CDSS tools were integrated into the ART clinic workflow, making it easy for the user to use it. The study has also revealed challenges in system reliability and information accuracy. Despite the challenges, the study further revealed that the system is effective and efficient, and overall, users are satisfied with the system. The study recommends that the implementers focus more on the logic behind the clinical decision-support intervention in order to address some of the concerns and enhance the accuracy of the information supplied. The study further suggests consulting the system's actual users throughout implementation.

Keywords: clinical decision support system, electronic medical record system, usability, antiretroviral therapy

Procedia PDF Downloads 109
6483 Water Body Detection and Estimation from Landsat Satellite Images Using Deep Learning

Authors: M. Devaki, K. B. Jayanthi

Abstract:

The identification of water bodies from satellite images has recently received a great deal of attention. Different methods have been developed to distinguish water bodies from various satellite images that vary in terms of time and space. Urban water identification issues body manifests in numerous applications with a great deal of certainty. There has been a sharp rise in the usage of satellite images to map natural resources, including urban water bodies and forests, during the past several years. This is because water and forest resources depend on each other so heavily that ongoing monitoring of both is essential to their sustainable management. The relevant elements from satellite pictures have been chosen using a variety of techniques, including machine learning. Then, a convolution neural network (CNN) architecture is created that can identify a superpixel as either one of two classes, one that includes water or doesn't from input data in a complex metropolitan scene. The deep learning technique, CNN, has advanced tremendously in a variety of visual-related tasks. CNN can improve classification performance by reducing the spectral-spatial regularities of the input data and extracting deep features hierarchically from raw pictures. Calculate the water body using the satellite image's resolution. Experimental results demonstrate that the suggested method outperformed conventional approaches in terms of water extraction accuracy from remote-sensing images, with an average overall accuracy of 97%.

Keywords: water body, Deep learning, satellite images, convolution neural network

Procedia PDF Downloads 93
6482 Determination of Optimum Conditions for the Leaching of Oxidized Copper Ores with Ammonium Nitrate

Authors: Javier Paul Montalvo Andia, Adriana Larrea Valdivia, Adolfo Pillihuaman Zambrano

Abstract:

The most common lixiviant in the leaching process of copper minerals is H₂SO₄, however, the current situation requires more environmentally friendly reagents and in certain situations that have a lower consumption due to the presence of undesirable gangue as muscovite or kaolinite that can make the process unfeasible. The present work studied the leaching of an oxidized copper mineral in an aqueous solution of ammonium nitrate, in order to obtain the optimum leaching conditions of the copper contained in the malachite mineral from Peru. The copper ore studied comes from a deposit in southern Peru and was characterized by X-ray diffractometer, inductively coupled-plasma emission spectrometer (ICP-OES) and atomic absorption spectrophotometry (AAS). The experiments were developed in batch reactor of 600 mL where the parameters as; temperature, pH, ammonium nitrate concentration, particle size and stirring speed were controlled according to experimental planning. The sample solution was analyzed for copper by atomic absorption spectrophotometry (AAS). A simulation in the HSC Chemistry 6.0 program showed that the predominance of the copper compounds of a Cu-H₂O aqueous system is altered by the presence in the system of ammonium complexes, the compound being thermodynamically more stable Cu(NH3)₄²⁺, which predominates in pH ranges from 8.5 to 10 at a temperature of 25 °C. The optimum conditions for copper leaching of the malachite mineral were a stirring speed of 600 rpm, an ammonium nitrate concentration of 4M, a particle diameter of 53 um and temperature of 62 °C. These results showed that the leaching of copper increases with increasing concentration of the ammonium solution, increasing the stirring rate, increasing the temperature and decreasing the particle diameter. Finally, the recovery of copper in optimum conditions was above 80%.

Keywords: ammonium nitrate, malachite, copper oxide, leaching

Procedia PDF Downloads 192
6481 MIMIC: A Multi Input Micro-Influencers Classifier

Authors: Simone Leonardi, Luca Ardito

Abstract:

Micro-influencers are effective elements in the marketing strategies of companies and institutions because of their capability to create an hyper-engaged audience around a specific topic of interest. In recent years, many scientific approaches and commercial tools have handled the task of detecting this type of social media users. These strategies adopt solutions ranging from rule based machine learning models to deep neural networks and graph analysis on text, images, and account information. This work compares the existing solutions and proposes an ensemble method to generalize them with different input data and social media platforms. The deployed solution combines deep learning models on unstructured data with statistical machine learning models on structured data. We retrieve both social media accounts information and multimedia posts on Twitter and Instagram. These data are mapped into feature vectors for an eXtreme Gradient Boosting (XGBoost) classifier. Sixty different topics have been analyzed to build a rule based gold standard dataset and to compare the performances of our approach against baseline classifiers. We prove the effectiveness of our work by comparing the accuracy, precision, recall, and f1 score of our model with different configurations and architectures. We obtained an accuracy of 0.91 with our best performing model.

Keywords: deep learning, gradient boosting, image processing, micro-influencers, NLP, social media

Procedia PDF Downloads 187
6480 Effect of Shot Peening on the Mechanical Properties for Welded Joints of Aluminium Alloy 6061-T6

Authors: Muna Khethier Abbass, Khairia Salman Hussan, Huda Mohummed AbdudAlaziz

Abstract:

This work aims to study the effect of shot peening on the mechanical properties of welded joints which performed by two different welding processes: Tungsten inert gas (TIG) welding and friction stir welding (FSW) processes of aluminum alloy 6061 T6. Arc welding process (TIG) was carried out on the sheet with dimensions of (100x50x6 mm) to obtain many welded joints with using electrode type ER4043 (AlSi5) as a filler metal and argon as shielding gas. While the friction stir welding process was carried out using CNC milling machine with a tool of rotational speed (1000 rpm) and welding speed of (20 mm/min) to obtain the same butt welded joints. The welded pieces were tested by X-ray radiography to detect the internal defects and faulty welded pieces were excluded. Tensile test specimens were prepared from welded joints and base alloy in the dimensions according to ASTM17500 and then subjected to shot peening process using steel ball of diameter 0.9 mm and for 15 min. All specimens were subjected to Vickers hardness test and micro structure examination to study the effect of welding process (TIG and FSW) on the micro structure of the weld zones. Results showed that a general decay of mechanical properties of TIG and FSW welded joints comparing with base alloy while the FSW welded joint gives better mechanical properties than that of TIG welded joint. This is due to the micro structure changes during the welding process. It has been found that the surface hardening by shot peening improved the mechanical properties of both welded joints, this is due to the compressive residual stress generation in the weld zones which was measured using X-Ray diffraction (XRD) inspection.

Keywords: friction stir welding, TIG welding, mechanical properties, shot peening

Procedia PDF Downloads 343
6479 A Static Android Malware Detection Based on Actual Used Permissions Combination and API Calls

Authors: Xiaoqing Wang, Junfeng Wang, Xiaolan Zhu

Abstract:

Android operating system has been recognized by most application developers because of its good open-source and compatibility, which enriches the categories of applications greatly. However, it has become the target of malware attackers due to the lack of strict security supervision mechanisms, which leads to the rapid growth of malware, thus bringing serious safety hazards to users. Therefore, it is critical to detect Android malware effectively. Generally, the permissions declared in the AndroidManifest.xml can reflect the function and behavior of the application to a large extent. Since current Android system has not any restrictions to the number of permissions that an application can request, developers tend to apply more than actually needed permissions in order to ensure the successful running of the application, which results in the abuse of permissions. However, some traditional detection methods only consider the requested permissions and ignore whether it is actually used, which leads to incorrect identification of some malwares. Therefore, a machine learning detection method based on the actually used permissions combination and API calls was put forward in this paper. Meanwhile, several experiments are conducted to evaluate our methodology. The result shows that it can detect unknown malware effectively with higher true positive rate and accuracy while maintaining a low false positive rate. Consequently, the AdaboostM1 (J48) classification algorithm based on information gain feature selection algorithm has the best detection result, which can achieve an accuracy of 99.8%, a true positive rate of 99.6% and a lowest false positive rate of 0.

Keywords: android, API Calls, machine learning, permissions combination

Procedia PDF Downloads 332
6478 Application of MALDI-MS to Differentiate SARS-CoV-2 and Non-SARS-CoV-2 Symptomatic Infections in the Early and Late Phases of the Pandemic

Authors: Dmitriy Babenko, Sergey Yegorov, Ilya Korshukov, Aidana Sultanbekova, Valentina Barkhanskaya, Tatiana Bashirova, Yerzhan Zhunusov, Yevgeniya Li, Viktoriya Parakhina, Svetlana Kolesnichenko, Yeldar Baiken, Aruzhan Pralieva, Zhibek Zhumadilova, Matthew S. Miller, Gonzalo H. Hortelano, Anar Turmuhambetova, Antonella E. Chesca, Irina Kadyrova

Abstract:

Introduction: The rapidly evolving COVID-19 pandemic, along with the re-emergence of pathogens causing acute respiratory infections (ARI), has necessitated the development of novel diagnostic tools to differentiate various causes of ARI. MALDI-MS, due to its wide usage and affordability, has been proposed as a potential instrument for diagnosing SARS-CoV-2 versus non-SARS-CoV-2 ARI. The aim of this study was to investigate the potential of MALDI-MS in conjunction with a machine learning model to accurately distinguish between symptomatic infections caused by SARS-CoV-2 and non-SARS-CoV-2 during both the early and later phases of the pandemic. Furthermore, this study aimed to analyze mass spectrometry (MS) data obtained from nasal swabs of healthy individuals. Methods: We gathered mass spectra from 252 samples, comprising 108 SARS-CoV-2-positive samples obtained in 2020 (Covid 2020), 7 SARS-CoV- 2-positive samples obtained in 2023 (Covid 2023), 71 samples from symptomatic individuals without SARS-CoV-2 (Control non-Covid ARVI), and 66 samples from healthy individuals (Control healthy). All the samples were subjected to RT-PCR testing. For data analysis, we employed the caret R package to train and test seven machine-learning algorithms: C5.0, KNN, NB, RF, SVM-L, SVM-R, and XGBoost. We conducted a training process using a five-fold (outer) nested repeated (five times) ten-fold (inner) cross-validation with a randomized stratified splitting approach. Results: In this study, we utilized the Covid 2020 dataset as a case group and the non-Covid ARVI dataset as a control group to train and test various machine learning (ML) models. Among these models, XGBoost and SVM-R demonstrated the highest performance, with accuracy values of 0.97 [0.93, 0.97] and 0.95 [0.95; 0.97], specificity values of 0.86 [0.71; 0.93] and 0.86 [0.79; 0.87], and sensitivity values of 0.984 [0.984; 1.000] and 1.000 [0.968; 1.000], respectively. When examining the Covid 2023 dataset, the Naive Bayes model achieved the highest classification accuracy of 43%, while XGBoost and SVM-R achieved accuracies of 14%. For the healthy control dataset, the accuracy of the models ranged from 0.27 [0.24; 0.32] for k-nearest neighbors to 0.44 [0.41; 0.45] for the Support Vector Machine with a radial basis function kernel. Conclusion: Therefore, ML models trained on MALDI MS of nasopharyngeal swabs obtained from patients with Covid during the initial phase of the pandemic, as well as symptomatic non-Covid individuals, showed excellent classification performance, which aligns with the results of previous studies. However, when applied to swabs from healthy individuals and a limited sample of patients with Covid in the late phase of the pandemic, ML models exhibited lower classification accuracy.

Keywords: SARS-CoV-2, MALDI-TOF MS, ML models, nasopharyngeal swabs, classification

Procedia PDF Downloads 115
6477 A Modularized Sensing Platform for Sensor Design Demonstration

Authors: Chun-Ming Huang, Yi-Jun Liu, Yi-Jie Hsieh, Jin-Ju Chue, Wei-Lin Lai, Chun-Yu Chen, Chih-Chyau Yang, Chien-Ming Wu

Abstract:

The market of wearable devices has been growing rapidly in two years. The integration of sensors and wearable devices has become the trend of the next technology products. Thus, the academics and industries are eager to cultivate talented persons in sensing technology. Currently, academic and industries have more and more demands on the integrations of versatile sensors and applications, especially for the teams who focus on the development of sensor circuit architectures. These teams tape-out many MEMs sensors chips through the chip fabrication service from National Chip Implementation Center (CIC). However, most of these teams are only able to focus on the circuit design of MEMs sensors; they lack the key support of further system demonstration. This paper follows the CIC’s main mission of promoting the chip/system advanced design technology and aims to establish the environments of the modularized sensing system platform and the system design flow with the measurement and calibration technology. These developed environments are used to support these research teams and help academically advanced sensor designs to perform the system demonstration. Thus, the research groups can promote and transfer their advanced sensor designs to industrial and further derive the industrial economic values. In this paper, the modularized sensing platform is proposed to enable the system demonstration for advanced sensor chip design. The environment of sensor measurement and calibration is established for academic to achieve an accurate sensor result. Two reference sensor designs cooperated with the modularized sensing platform are given to show the sensing system integration and demonstration. These developed environments and platforms are currently provided to academics in Taiwan, and so that the academics can obtain a better environment to perform the system demonstration and improve the research and teaching quality.

Keywords: modularized sensing platform, sensor design and calibration, sensor system, sensor system design flow

Procedia PDF Downloads 239
6476 Evaluation of Commercial Back-analysis Package in Condition Assessment of Railways

Authors: Shadi Fathi, Moura Mehravar, Mujib Rahman

Abstract:

Over the years,increased demands on railways, the emergence of high-speed trains and heavy axle loads, ageing, and deterioration of the existing tracks, is imposing costly maintenance actions on the railway sector. The need for developing a fast andcost-efficient non-destructive assessment method for the structural evaluation of railway tracksis therefore critically important. The layer modulus is the main parameter used in the structural design and evaluation of the railway track substructure (foundation). Among many recently developed NDTs, Falling Weight Deflectometer (FWD) test, widely used in pavement evaluation, has shown promising results for railway track substructure monitoring. The surface deflection data collected by FWD are used to estimate the modulus of substructure layers through the back-analysis technique. Although there are different commerciallyavailableback-analysis programs are used for pavement applications, there are onlya limited number of research-based techniques have been so far developed for railway track evaluation. In this paper, the suitability, accuracy, and reliability of the BAKFAAsoftware are investigated. The main rationale for selecting BAKFAA as it has a relatively straightforward user interfacethat is freely available and widely used in highway and airport pavement evaluation. As part of the study, a finite element (FE) model of a railway track section near Leominsterstation, Herefordshire, UK subjected to the FWD test, was developed and validated against available field data. Then, a virtual experimental database (including 218 sets of FWD testing data) was generated using theFE model and employed as the measured database for the BAKFAA software. This database was generated considering various layers’ moduli for each layer of track substructure over a predefined range. The BAKFAA predictions were compared against the cone penetration test (CPT) data (available from literature; conducted near to Leominster station same section as the FWD was performed). The results reveal that BAKFAA overestimatesthe layers’ moduli of each substructure layer. To adjust the BAKFA with the CPT data, this study introduces a correlation model to make the BAKFAA applicable in railway applications.

Keywords: back-analysis, bakfaa, railway track substructure, falling weight deflectometer (FWD), cone penetration test (CPT)

Procedia PDF Downloads 133
6475 Percentage Contribution of Lower Limb Moments to Vertical Ground Reaction Force in Normal Walking

Authors: Salam M. Elhafez, Ahmed A. Ashour, Naglaa M. Elhafez, Ghada M. Elhafez, Azza M. Abdelmohsen

Abstract:

Patients suffering from gait disturbances are referred by having muscle group dysfunctions. There is a need for more studies investigating the contribution of muscle moments of the lower limb to the vertical ground reaction force using 3D gait analysis system. The purpose of this study was to investigate how the hip, knee and ankle moments in the sagittal plane contribute to the vertical ground reaction force in healthy subjects during normal speed of walking. Forty healthy male individuals volunteered to participate in this study. They were filmed using six high speed (120 Hz) Pro-Reflex Infrared cameras (Qualisys) while walking on an AMTI force platform. The data collected were the percentage contribution of the moments of the hip, knee and ankle joints in the sagittal plane at the instant of occurrence of the first peak, second peak, and the trough of the vertical ground reaction force. The results revealed that at the first peak of the ground reaction force (loading response), the highest contribution was generated from the knee extension moment, followed by the hip extension moment. Knee flexion and ankle plantar flexion moments produced high contribution to the trough of the ground reaction force (midstance) with approximately equal values. The second peak of the ground reaction force was mainly produced by the ankle plantar flexion moment. Conclusion: Hip and knee flexion and extension moments and ankle plantar flexion moment play important roles in the supporting phase of normal walking.

Keywords: gait analysis, ground reaction force, moment contribution, normal walking

Procedia PDF Downloads 381
6474 Provision of Afterschool Programs: Understanding the Educational Needs and Outcomes of Newcomer and Refugee Students in Canada

Authors: Edward Shizha, Edward Makwarimba

Abstract:

Newcomer and refugee youth feel excluded in the education system in Canada, and the formal education environment does not fully cater for their learning needs. The objective of this study was to build knowledge and understanding of the educational needs and experiences of these youth in Canada and how available afterschool programs can most effectively support their learning needs and academic outcomes. The Employment and Social Development Canada (ESDC), which funded this research, enables and empowers students to advance their educational experience through targeted investments in services that are delivered by youth-serving organizations outside the formal education system through afterschool initiatives. A literature review and a provincial/territorial internet scan were conducted to determine the availability of services and programs that serve the educational needs and academic outcomes of newcomer youth in 10 provinces and 3 territories in Canada. The goal was to identify intersectional factors (e.g., gender, sexuality, culture, social class, race, etc.) that influence educational outcomes of newcomer/refugee students and to recommend ways the ESDC could complement settlement services to enhance students’ educational success. First, data was collected through a literature search of various databases, including PubMed, Web of Science, Scopus, Google docs, ACADEMIA, and grey literature, including government documents, to inform our analysis. Second, a provincial/territorial internet scan was conducted using a template that was created by ESDC staff with the input of the researchers. The objective of the web-search scan was to identify afterschool programs, projects, and initiatives offered to newcomer/refugee youth by service provider organizations. The method for the scan included both qualitative and quantitative data gathering. Both the literature review and the provincial/territorial scan revealed that there are gender disparities in educational outcomes of newcomer and refugee youth. High school completion rates by gender show that boys are at higher risk of not graduating than girls and that girls are more likely than boys to have at least a high school diploma and more likely to proceed to postsecondary education. Findings from literature reveal that afterschool programs are required for refugee youth who experience mental health challenges and miss out on significant periods of schooling, which affect attendance, participation, and graduation from high school. However, some refugee youth use their resilience and ambition to succeed in their educational outcomes. Another finding showed that some immigrant/refugee students, through ethnic organizations and familial affiliation, maintain aspects of their cultural values, parental expectations and ambitious expectations for their own careers to succeed in both high school and postsecondary education. The study found a significant combination of afterschool programs that include academic support, scholarships, bursaries, homework support, career readiness, internships, mentorship, tutoring, non-clinical counselling, mental health and social well-being support, language skills, volunteering opportunities, community connections, peer networking, culturally relevant services etc. These programs assist newcomer youth to develop self-confidence and prepare for academic success and future career development. The study concluded that advantages of afterschool programs are greatest for youth at risk for poor educational outcomes, such as Latino and Black youth, including 2SLGBTQI+ immigrant youth.

Keywords: afterschool programs, educational outcomes, newcomer youth, refugee youth, youth-serving organizations

Procedia PDF Downloads 82
6473 Challenges and Success Factors in Introducing Information Systems for Students' Online Registration

Authors: Stanley Fore, Sharon Chipeperekwa

Abstract:

The start of the 2011 academic year in South Africa saw a number of Institutions of Higher Learning introducing online registration for their students. The efficiency and effectiveness of Information Systems are increasingly becoming a necessity and not an option for many organizations. An information system should be able to allow end users to access information easily and navigate with ease. The selected University of Technology (UoT) in this research is one of the largest public institution of higher learning in the Western Cape Province and boasts of an enrolment of more than 30000 students per academic year. An observation was made that, during registration students’ stand in long queues waiting to register or for assistance to register. The system tends to ‘freeze’ whilst students are registering and students are in most cases unfamiliar with the system interface. They constantly have to enquire what to do next when going through online registration process. A mixed method approach will be adopted which comprises of quantitative and qualitative approaches. The study uses constructs of the updated DeLone and McLean IS success model (2003) to analyse and explain the student’s perceptions of the online registration system. The research was undertaken to establish the student’s perceptions of the online registration system. This research seeks to identify and analyse the challenges and success factors of introducing an online registration system whilst highlighting the extent to which this system has been able to solve the numerous problems associated with the manual era. The study will assist management and those responsible for managing the current system to determine how well the system is working or not working to achieve user satisfaction. It will also assist them going forward on what to consider before, during and after implementation of an information system. Respondents will be informed of the objectives of the research, and their consent to participate will be sought. Ethical considerations that will be applied to this study include; informed consent and protection from harm, right to privacy and involvement of the research.

Keywords: online registration, information systems, University of Technology, end-users

Procedia PDF Downloads 266
6472 Validation of Asymptotic Techniques to Predict Bistatic Radar Cross Section

Authors: M. Pienaar, J. W. Odendaal, J. C. Smit, J. Joubert

Abstract:

Simulations are commonly used to predict the bistatic radar cross section (RCS) of military targets since characterization measurements can be expensive and time consuming. It is thus important to accurately predict the bistatic RCS of targets. Computational electromagnetic (CEM) methods can be used for bistatic RCS prediction. CEM methods are divided into full-wave and asymptotic methods. Full-wave methods are numerical approximations to the exact solution of Maxwell’s equations. These methods are very accurate but are computationally very intensive and time consuming. Asymptotic techniques make simplifying assumptions in solving Maxwell's equations and are thus less accurate but require less computational resources and time. Asymptotic techniques can thus be very valuable for the prediction of bistatic RCS of electrically large targets, due to the decreased computational requirements. This study extends previous work by validating the accuracy of asymptotic techniques to predict bistatic RCS through comparison with full-wave simulations as well as measurements. Validation is done with canonical structures as well as complex realistic aircraft models instead of only looking at a complex slicy structure. The slicy structure is a combination of canonical structures, including cylinders, corner reflectors and cubes. Validation is done over large bistatic angles and at different polarizations. Bistatic RCS measurements were conducted in a compact range, at the University of Pretoria, South Africa. The measurements were performed at different polarizations from 2 GHz to 6 GHz. Fixed bistatic angles of β = 30.8°, 45° and 90° were used. The measurements were calibrated with an active calibration target. The EM simulation tool FEKO was used to generate simulated results. The full-wave multi-level fast multipole method (MLFMM) simulated results together with the measured data were used as reference for validation. The accuracy of physical optics (PO) and geometrical optics (GO) was investigated. Differences relating to amplitude, lobing structure and null positions were observed between the asymptotic, full-wave and measured data. PO and GO were more accurate at angles close to the specular scattering directions and the accuracy seemed to decrease as the bistatic angle increased. At large bistatic angles PO did not perform well due to the shadow regions not being treated appropriately. PO also did not perform well for canonical structures where multi-bounce was the main scattering mechanism. PO and GO do not account for diffraction but these inaccuracies tended to decrease as the electrical size of objects increased. It was evident that both asymptotic techniques do not properly account for bistatic structural shadowing. Specular scattering was calculated accurately even if targets did not meet the electrically large criteria. It was evident that the bistatic RCS prediction performance of PO and GO depends on incident angle, frequency, target shape and observation angle. The improved computational efficiency of the asymptotic solvers yields a major advantage over full-wave solvers and measurements; however, there is still much room for improvement of the accuracy of these asymptotic techniques.

Keywords: asymptotic techniques, bistatic RCS, geometrical optics, physical optics

Procedia PDF Downloads 261
6471 Modeling Fertility and Production of Hazelnut Cultivars through the Artificial Neural Network under Climate Change of Karaj

Authors: Marziyeh Khavari

Abstract:

In recent decades, climate change, global warming, and the growing population worldwide face some challenges, such as increasing food consumption and shortage of resources. Assessing how climate change could disturb crops, especially hazelnut production, seems crucial for sustainable agriculture production. For hazelnut cultivation in the mid-warm condition, such as in Iran, here we present an investigation of climate parameters and how much they are effective on fertility and nut production of hazelnut trees. Therefore, the climate change of the northern zones in Iran has investigated (1960-2017) and was reached an uptrend in temperature. Furthermore, the descriptive analysis performed on six cultivars during seven years shows how this small-scale survey could demonstrate the effects of climate change on hazelnut production and stability. Results showed that some climate parameters are more significant on nut production, such as solar radiation, soil temperature, relative humidity, and precipitation. Moreover, some cultivars have produced more stable production, for instance, Negret and Segorbe, while the Mervill de Boliver recorded the most variation during the study. Another aspect that needs to be met is training and predicting an actual model to simulate nut production through a neural network and linear regression simulation. The study developed and estimated the ANN model's generalization capability with different criteria such as RMSE, SSE, and accuracy factors for dependent and independent variables (environmental and yield traits). The models were trained and tested while the accuracy of the model is proper to predict hazelnut production under fluctuations in weather parameters.

Keywords: climate change, neural network, hazelnut, global warming

Procedia PDF Downloads 136
6470 A Methodology for Automatic Diversification of Document Categories

Authors: Dasom Kim, Chen Liu, Myungsu Lim, Su-Hyeon Jeon, ByeoungKug Jeon, Kee-Young Kwahk, Namgyu Kim

Abstract:

Recently, numerous documents including unstructured data and text have been created due to the rapid increase in the usage of social media and the Internet. Each document is usually provided with a specific category for the convenience of the users. In the past, the categorization was performed manually. However, in the case of manual categorization, not only can the accuracy of the categorization be not guaranteed but the categorization also requires a large amount of time and huge costs. Many studies have been conducted towards the automatic creation of categories to solve the limitations of manual categorization. Unfortunately, most of these methods cannot be applied to categorizing complex documents with multiple topics because the methods work by assuming that one document can be categorized into one category only. In order to overcome this limitation, some studies have attempted to categorize each document into multiple categories. However, they are also limited in that their learning process involves training using a multi-categorized document set. These methods therefore cannot be applied to multi-categorization of most documents unless multi-categorized training sets are provided. To overcome the limitation of the requirement of a multi-categorized training set by traditional multi-categorization algorithms, we previously proposed a new methodology that can extend a category of a single-categorized document to multiple categorizes by analyzing relationships among categories, topics, and documents. In this paper, we design a survey-based verification scenario for estimating the accuracy of our automatic categorization methodology.

Keywords: big data analysis, document classification, multi-category, text mining, topic analysis

Procedia PDF Downloads 276
6469 Investigating the Sloshing Characteristics of a Liquid by Using an Image Processing Method

Authors: Ufuk Tosun, Reza Aghazadeh, Mehmet Bülent Özer

Abstract:

This study puts forward a method to analyze the sloshing characteristics of liquid in a tuned sloshing absorber system by using image processing tools. Tuned sloshing vibration absorbers have recently attracted researchers’ attention as a seismic load damper in constructions due to its practical and logistical convenience. The absorber is liquid which sloshes and applies a force in opposite phase to the motion of structure. Experimentally characterization of the sloshing behavior can be utilized as means of verifying the results of numerical analysis. It can also be used to identify the accuracy of assumptions related to the motion of the liquid. There are extensive theoretical and experimental studies in the literature related to the dynamical and structural behavior of tuned sloshing dampers. In most of these works there are efforts to estimate the sloshing behavior of the liquid such as free surface motion and total force applied by liquid to the wall of container. For these purposes the use of sensors such as load cells and ultrasonic sensors are prevalent in experimental works. Load cells are only capable of measuring the force and requires conducting tests both with and without liquid to obtain pure sloshing force. Ultrasonic level sensors give point-wise measurements and hence they are not applicable to measure the whole free surface motion. Furthermore, in the case of liquid splashing it may give incorrect data. In this work a method for evaluating the sloshing wave height by using camera records and image processing techniques is presented. In this method the motion of the liquid and its container, made of a transparent material, is recorded by a high speed camera which is aligned to the free surface of the liquid. The video captured by the camera is processed frame by frame by using MATLAB Image Processing toolbox. The process starts with cropping the desired region. By recognizing the regions containing liquid and eliminating noise and liquid splashing, the final picture depicting the free surface of liquid is achieved. This picture then is used to obtain the height of the liquid through the length of container. This process is verified by ultrasonic sensors that measured fluid height on the surface of liquid.

Keywords: fluid structure interaction, image processing, sloshing, tuned liquid damper

Procedia PDF Downloads 347
6468 Nonviolent Communication and Disciplinary Area of Social Communication: Case Study on the International Circulation of Ideas from a Brazilian Perspective

Authors: Luiza Toschi

Abstract:

This work presents part of an empirical and theoretical master's degree meta-research that is interested in the relationship between the disciplinary area of Social Communication, to be investigated with the characteristics of the Bourdieusian scientific field, and the emergence of public interest in Nonviolent Communication (NVC) in Brazil and the world. To this end, the state of the art of this conceptual and practical relationship is investigated based on scientific productions available in spaces of academic credibility, such as conferences and scientific journals renowned in the field. From there, agents and the sociological aspects that make them contribute or not to scientific production in Brazil and the world are mapped. In this work, a brief dive into the international context is presented to understand if and how nonviolent communication permeates scientific production in communication in a systematic way. Using three accessible articles published between 2013 and 2022 in the 117 magazines classified as Quartiles Q1 in the Journal Ranking of Communication, the international production on the subject is compared with the Brazilian one from its context. The social conditions of the international circulation of ideas are thus discussed. Science is a product of its social environment, arising from relations of interest and power that compete in the political dimension at the same time as in the epistemological dimension. In this way, scientific choices are linked to the resources mobilized from or through the prestige and recognition of peers. In this sense, an object of interest stands out to a scientist for its academic value, but also and inseparably that which has a social interest within the collective, their social stratification, and the context of legitimacy created in their surroundings, influenced by cultural universalism. In Brazil, three published articles were found in congresses and journals that mention NVC in their abstract or keywords. All were written by Public Relations undergraduate students. Between the most experienced researchers who guided or validated the publications, it is possible to find two professionals who are interested in the Culture of Peace and Dialogy. Likewise, internationally, only three of the articles found mention the term in their abstract or title. Two analyze journalistic coverage based on the principles of NVC and Journalism for Peace. The third is from one of the Brazilian researchers identified as interested in dialogic practices, who analyses audiovisual material and promotes epistemological reflections. If, on the one hand, some characteristics inside and outside Brazil are similar: small samples, relationship with peace studies, and female researchers, two of whom are Brazilian, on the other hand, differences are obvious. If within the country, the subject is mostly Organizational Communication, outside this intersection, it is not presented explicitly. Furthermore, internationally, there is an interest in analyzing from the perspective of NVC, which has not been found so far in publications in Brazil. Up to the present moment, it is possible to presume that, universally, the legitimacy of the topic is sought by its association with conflict conciliation research and communication for peace.

Keywords: academic field sociology, international circulation of ideas, meta research in communication, nonviolent communication

Procedia PDF Downloads 44
6467 Enhanced Planar Pattern Tracking for an Outdoor Augmented Reality System

Authors: L. Yu, W. K. Li, S. K. Ong, A. Y. C. Nee

Abstract:

In this paper, a scalable augmented reality framework for handheld devices is presented. The presented framework is enabled by using a server-client data communication structure, in which the search for tracking targets among a database of images is performed on the server-side while pixel-wise 3D tracking is performed on the client-side, which, in this case, is a handheld mobile device. Image search on the server-side adopts a residual-enhanced image descriptors representation that gives the framework a scalability property. The tracking algorithm on the client-side is based on a gravity-aligned feature descriptor which takes the advantage of a sensor-equipped mobile device and an optimized intensity-based image alignment approach that ensures the accuracy of 3D tracking. Automatic content streaming is achieved by using a key-frame selection algorithm, client working phase monitoring and standardized rules for content communication between the server and client. The recognition accuracy test performed on a standard dataset shows that the method adopted in the presented framework outperforms the Bag-of-Words (BoW) method that has been used in some of the previous systems. Experimental test conducted on a set of video sequences indicated the real-time performance of the tracking system with a frame rate at 15-30 frames per second. The presented framework is exposed to be functional in practical situations with a demonstration application on a campus walk-around.

Keywords: augmented reality framework, server-client model, vision-based tracking, image search

Procedia PDF Downloads 279
6466 Fast Approximate Bayesian Contextual Cold Start Learning (FAB-COST)

Authors: Jack R. McKenzie, Peter A. Appleby, Thomas House, Neil Walton

Abstract:

Cold-start is a notoriously difficult problem which can occur in recommendation systems, and arises when there is insufficient information to draw inferences for users or items. To address this challenge, a contextual bandit algorithm – the Fast Approximate Bayesian Contextual Cold Start Learning algorithm (FAB-COST) – is proposed, which is designed to provide improved accuracy compared to the traditionally used Laplace approximation in the logistic contextual bandit, while controlling both algorithmic complexity and computational cost. To this end, FAB-COST uses a combination of two moment projection variational methods: Expectation Propagation (EP), which performs well at the cold start, but becomes slow as the amount of data increases; and Assumed Density Filtering (ADF), which has slower growth of computational cost with data size but requires more data to obtain an acceptable level of accuracy. By switching from EP to ADF when the dataset becomes large, it is able to exploit their complementary strengths. The empirical justification for FAB-COST is presented, and systematically compared to other approaches on simulated data. In a benchmark against the Laplace approximation on real data consisting of over 670, 000 impressions from autotrader.co.uk, FAB-COST demonstrates at one point increase of over 16% in user clicks. On the basis of these results, it is argued that FAB-COST is likely to be an attractive approach to cold-start recommendation systems in a variety of contexts.

Keywords: cold-start learning, expectation propagation, multi-armed bandits, Thompson Sampling, variational inference

Procedia PDF Downloads 111
6465 Procedural Protocol for Dual Energy Computed Tomography (DECT) Inversion

Authors: Rezvan Ravanfar Haghighi, S. Chatterjee, Pratik Kumar, V. C. Vani, Priya Jagia, Sanjiv Sharma, Susama Rani Mandal, R. Lakshmy

Abstract:

The dual energy computed tomography (DECT) aims at noting the HU(V) values for the sample at two different voltages V=V1, V2 and thus obtain the electron densities (ρe) and effective atomic number (Zeff) of the substance. In the present paper, we aim to obtain a numerical algorithm by which (ρe, Zeff) can be obtained from the HU(100) and HU(140) data, where V=100, 140 kVp. The idea is to use this inversion method to characterize and distinguish between the lipid and fibrous coronary artery plaques.With the idea to develop the inversion algorithm for low Zeff materials, as is the case with non calcified coronary artery plaque, we prepare aqueous samples whose calculated values of (ρe, Zeff) lie in the range (2.65×1023≤ ρe≤ 3.64×1023 per cc ) and (6.80≤ Zeff ≤ 8.90). We fill the phantom with these known samples and experimentally determine HU(100) and HU(140) for the same pixels. Knowing that the HU(V) values are related to the attenuation coefficient of the system, we present an algorithm by which the (ρe, Zeff) is calibrated with respect to (HU(100), HU(140)). The calibration is done with a known set of 20 samples; its accuracy is checked with a different set of 23 known samples. We find that the calibration gives the ρe with an accuracy of ± 4% while Zeff is found within ±1% of the actual value, the confidence being 95%.In this inversion method (ρe, Zeff) of the scanned sample can be found by eliminating the effects of the CT machine and also by ensuring that the determination of the two unknowns (ρe, Zeff) does not interfere with each other. It is found that this algorithm can be used for prediction of chemical characteristic (ρe, Zeff) of unknown scanned materials with 95% confidence level, by inversion of the DECT data.

Keywords: chemical composition, dual-energy computed tomography, inversion algorithm

Procedia PDF Downloads 440
6464 Interpretation of the Russia-Ukraine 2022 War via N-Gram Analysis

Authors: Elcin Timur Cakmak, Ayse Oguzlar

Abstract:

This study presents the results of the tweets sent by Twitter users on social media about the Russia-Ukraine war by bigram and trigram methods. On February 24, 2022, Russian President Vladimir Putin declared a military operation against Ukraine, and all eyes were turned to this war. Many people living in Russia and Ukraine reacted to this war and protested and also expressed their deep concern about this war as they felt the safety of their families and their futures were at stake. Most people, especially those living in Russia and Ukraine, express their views on the war in different ways. The most popular way to do this is through social media. Many people prefer to convey their feelings using Twitter, one of the most frequently used social media tools. Since the beginning of the war, it is seen that there have been thousands of tweets about the war from many countries of the world on Twitter. These tweets accumulated in data sources are extracted using various codes for analysis through Twitter API and analysed by Python programming language. The aim of the study is to find the word sequences in these tweets by the n-gram method, which is known for its widespread use in computational linguistics and natural language processing. The tweet language used in the study is English. The data set consists of the data obtained from Twitter between February 24, 2022, and April 24, 2022. The tweets obtained from Twitter using the #ukraine, #russia, #war, #putin, #zelensky hashtags together were captured as raw data, and the remaining tweets were included in the analysis stage after they were cleaned through the preprocessing stage. In the data analysis part, the sentiments are found to present what people send as a message about the war on Twitter. Regarding this, negative messages make up the majority of all the tweets as a ratio of %63,6. Furthermore, the most frequently used bigram and trigram word groups are found. Regarding the results, the most frequently used word groups are “he, is”, “I, do”, “I, am” for bigrams. Also, the most frequently used word groups are “I, do, not”, “I, am, not”, “I, can, not” for trigrams. In the machine learning phase, the accuracy of classifications is measured by Classification and Regression Trees (CART) and Naïve Bayes (NB) algorithms. The algorithms are used separately for bigrams and trigrams. We gained the highest accuracy and F-measure values by the NB algorithm and the highest precision and recall values by the CART algorithm for bigrams. On the other hand, the highest values for accuracy, precision, and F-measure values are achieved by the CART algorithm, and the highest value for the recall is gained by NB for trigrams.

Keywords: classification algorithms, machine learning, sentiment analysis, Twitter

Procedia PDF Downloads 79
6463 An Exploratory Study on the Integration of Neurodiverse University Students into Mainstream Learning and Their Performance: The Case of the Jones Learning Center

Authors: George Kassar, Phillip A. Cartwright

Abstract:

Based on data collected from The Jones Learning Center (JLC), University of the Ozarks, Arkansas, U.S., this study explores the impact of inclusive classroom practices on neuro-diverse college students’ and their consequent academic performance having participated in integrative therapies designed to support students who are intellectually capable of obtaining a college degree, but who require support for learning challenges owing to disabilities, AD/HD, or ASD. The purpose of this study is two-fold. The first objective is to explore the general process, special techniques, and practices of the (JLC) inclusive program. The second objective is to identify and analyze the effectiveness of the processes, techniques, and practices in supporting the academic performance of enrolled college students with learning disabilities following integration into mainstream university learning. Integrity, transparency, and confidentiality are vital in the research. All questions were shared in advance and confirmed by the concerned management at the JLC. While administering the questionnaire as well as conducted the interviews, the purpose of the study, its scope, aims, and objectives were clearly explained to all participants prior starting the questionnaire / interview. Confidentiality of all participants assured and guaranteed by using encrypted identification of individuals, thus limiting access to data to only the researcher, and storing data in a secure location. Respondents were also informed that their participation in this research is voluntary, and they may withdraw from it at any time prior to submission if they wish. Ethical consent was obtained from the participants before proceeding with videorecording of the interviews. This research uses a mixed methods approach. The research design involves collecting, analyzing, and “mixing” quantitative and qualitative methods and data to enable a research inquiry. The research process is organized based on a five-pillar approach. The first three pillars are focused on testing the first hypothesis (H1) directed toward determining the extent to the academic performance of JLC students did improve after involvement with comprehensive JLC special program. The other two pillars relate to the second hypothesis (H2), which is directed toward determining the extent to which collective and applied knowledge at JLC is distinctive from typical practices in the field. The data collected for research were obtained from three sources: 1) a set of secondary data in the form of Grade Point Average (GPA) received from the registrar, 2) a set of primary data collected throughout structured questionnaire administered to students and alumni at JLC, and 3) another set of primary data collected throughout interviews conducted with staff and educators at JLC. The significance of this study is two folds. First, it validates the effectiveness of the special program at JLC for college-level students who learn differently. Second, it identifies the distinctiveness of the mix of techniques, methods, and practices, including the special individualized and personalized one-on-one approach at JLC.

Keywords: education, neuro-diverse students, program effectiveness, Jones learning center

Procedia PDF Downloads 77
6462 A Study on the Effectiveness of Alternative Commercial Ventilation Inlets That Improve Energy Efficiency of Building Ventilation Systems

Authors: Brian Considine, Aonghus McNabola, John Gallagher, Prashant Kumar

Abstract:

Passive air pollution control devices known as aspiration efficiency reducers (AER) have been developed using aspiration efficiency (AE) concepts. Their purpose is to reduce the concentration of particulate matter (PM) drawn into a building air handling unit (AHU) through alterations in the inlet design improving energy consumption. In this paper an examination is conducted into the effect of installing a deflector system around an AER-AHU inlet for both a forward and rear-facing orientations relative to the wind. The results of the study found that these deflectors are an effective passive control method for reducing AE at various ambient wind speeds over a range of microparticles of varying diameter. The deflector system was found to induce a large wake zone at low ambient wind speeds for a rear-facing AER-AHU, resulting in significantly lower AE in comparison to without. As the wind speed increased, both contained a wake zone but have much lower concentration gradients with the deflectors. For the forward-facing models, the deflector system at low ambient wind speed was preferred at higher Stokes numbers but there was negligible difference as the Stokes number decreased. Similarly, there was no significant difference at higher wind speeds across the Stokes number range tested. The results demonstrate that a deflector system is a viable passive control method for the reduction of ventilation energy consumption.

Keywords: air handling unit, air pollution, aspiration efficiency, energy efficiency, particulate matter, ventilation

Procedia PDF Downloads 123
6461 Effective Energy Saving of a Large Building through Multiple Approaches

Authors: Choo Hong Ang

Abstract:

The most popular approach to save energy for large commercial buildings in Malaysia is to replace the existing chiller plant of high kW/ton to one of lower kW/ton. This approach, however, entails large capital outlay with a long payment period of up to 7 years. This paper shows that by using multiple approaches, other than replacing the existing chiller plant, an energy saving of up to 20 %, is possible. The main methodology adopted was to identify and then plugged all heat ingress paths into a building, including putting up glass structures to prevent mixing of internal air-conditioned air with the ambient environment, and replacing air curtains with glass doors. This methodology could save up to 10 % energy bill. Another methodology was to change fixed speed motors of air handling units (AHU) to variable speed drive (VSD) and changing escalators to motion-sensor type. Other methodologies included reducing heat load by blocking air supply to non-occupied parcels, rescheduling chiller plant operation, changing of fluorescent lights to LED lights, and conversion from tariff B to C1. A case example of Komtar, the tallest building in Penang, is given here. The total energy bill for Komtar was USD2,303,341 in 2016 but was reduced to USD 1,842,927.39 in 2018, a significant saving of USD460,413.86 or 20 %. In terms of kWh, there was a reduction from 18, 302,204.00 kWh in 2016 to 14,877,105.00 kWh in 2018, a reduction of 3,425,099.00 kWh or 18.71 %. These methodologies used were relatively low cost and the payback period was merely 24 months. With this achievement, the Komtar building was awarded champion of the Malaysian National Energy Award 2019 and second runner up of the Asean Energy Award. This experience shows that a strong commitment to energy saving is the key to effective energy saving.

Keywords: chiller plant, energy saving measures, heat ingress, large building

Procedia PDF Downloads 109