Search results for: cloud effective radius
10029 Thermodynamics of Aqueous Solutions of Organic Molecule and Electrolyte: Use Cloud Point to Obtain Better Estimates of Thermodynamic Parameters
Authors: Jyoti Sahu, Vinay A. Juvekar
Abstract:
Electrolytes are often used to bring about salting-in and salting-out of organic molecules and polymers (e.g. polyethylene glycols/proteins) from the aqueous solutions. For quantification of these phenomena, a thermodynamic model which can accurately predict activity coefficient of electrolyte as a function of temperature is needed. The thermodynamics models available in the literature contain a large number of empirical parameters. These parameters are estimated using lower/upper critical solution temperature of the solution in the electrolyte/organic molecule at different temperatures. Since the number of parameters is large, inaccuracy can bethe creep in during their estimation, which can affect the reliability of prediction beyond the range in which these parameters are estimated. Cloud point of solution is related to its free energy through temperature and composition derivative. Hence, the Cloud point measurement can be used for accurate estimation of the temperature and composition dependence of parameters in the model for free energy. Hence, if we use a two pronged procedure in which we first use cloud point of solution to estimate some of the parameters of the thermodynamic model and determine the rest using osmotic coefficient data, we gain on two counts. First, since the parameters, estimated in each of the two steps, are fewer, we achieve higher accuracy of estimation. The second and more important gain is that the resulting model parameters are more sensitive to temperature. This is crucial when we wish to use the model outside temperatures window within which the parameter estimation is sought. The focus of the present work is to prove this proposition. We have used electrolyte (NaCl/Na2CO3)-water-organic molecule (Iso-propanol/ethanol) as the model system. The model of Robinson-Stokes-Glukauf is modified by incorporating the temperature dependent Flory-Huggins interaction parameters. The Helmholtz free energy expression contains, in addition to electrostatic and translational entropic contributions, three Flory-Huggins pairwise interaction contributions viz., and (w-water, p-polymer, s-salt). These parameters depend both on temperature and concentrations. The concentration dependence is expressed in the form of a quadratic expression involving the volume fractions of the interacting species. The temperature dependence is expressed in the form .To obtain the temperature-dependent interaction parameters for organic molecule-water and electrolyte-water systems, Critical solution temperature of electrolyte -water-organic molecules is measured using cloud point measuring apparatus The temperature and composition dependent interaction parameters for electrolyte-water-organic molecule are estimated through measurement of cloud point of solution. The model is used to estimate critical solution temperature (CST) of electrolyte water-organic molecules solution. We have experimentally determined the critical solution temperature of different compositions of electrolyte-water-organic molecule solution and compared the results with the estimates based on our model. The two sets of values show good agreement. On the other hand when only osmotic coefficients are used for estimation of the free energy model, CST predicted using the resulting model show poor agreement with the experiments. Thus, the importance of the CST data in the estimation of parameters of the thermodynamic model is confirmed through this work.Keywords: concentrated electrolytes, Debye-Hückel theory, interaction parameters, Robinson-Stokes-Glueckauf model, Flory-Huggins model, critical solution temperature
Procedia PDF Downloads 39110028 Effect of Geometry on the Aerodynamic Performance of Darrieus H Yype Vertical Axis Wind Turbine
Authors: Belkheir Noura, Rabah Kerfah, Boumehani Abdellah
Abstract:
The influence of solidity variations on the aerodynamic performance of H type vertical axis wind turbine is studied in this paper. The wind turbine model used in this paper is the three-blade wind turbine with the symmetrical airfoil, NACA0021. The length of the chord is 0.265m. Numerical investigations were implemented for the different solidity by changing the radius and blade number. A two-dimensional model of the wind turbine is employed. The approach a Reynolds-Averaged Navier–Stokes equations, completed by the K- ώ SST turbulence model, is used. Motion mesh model capability of a computational fluid dynamics (CFD) solver is used. For each value of the solidity, the aerodynamics performances and the characteristics of the flow field are studied at several values of the tip speed ratio, λ = 0.5 to λ = 3, with an incoming wind speed of 8 m/s. The results show that increasing the number of blades will reduce the maximum value of the power coefficient of the wind turbine. Also, for the VAWT with a lower solidity can obtain the maximum Cp at a high tip speed ratio. The effects of changing the radius and blade number on aerodynamic performance are almost the same. Finally, for the validation, experimental data from the literature and computational results were compared. In conclusion, to study the influence of the solidity in the performances of the wind turbine is to provide the reference for the design of H type vertical axis wind turbines.Keywords: wind energy, darrieus h type vertical axis wind turbine, computational fluid dynamic, solidity
Procedia PDF Downloads 9510027 Analyzing the Quality of Cloud-Based E-Learning Systems on the Perception of the Learners and the Teachers
Authors: R. W. C. Devindi, S. M. Buddika Harshanath
Abstract:
E-learning is a widely used technology for learning in the modern world. With the pandemic situation the popularity of using e-learning has been increased in a larger capacity. The e-learning educational systems require software resources as well as hardware usually but it is hard for most of the education institutions to afford those resources. Also with the massive user load e-learning has to broaden the server side resources as well. Therefore, in the present cloud computing was implemented in order to make the e – learning systems more efficient. The researcher has analyzed the quality of the e-learning systems on the perception of the learners and the teachers with the aid of hypothesis and has given the analyzed results and the discussion in this report. Therefore, the future research will be able to get some steps to increase the quality of the online learning systems furthermore. In the case of e-learning, quality assurance and cost effectiveness are essential. A complex quality assurance system is used in the stated project. There are no well-defined standard evaluation measures in this field. As a result, accurately assessing the e-learning system's overall quality is challenging. The researcher has done the analysis with the aid of standard methods and software.Keywords: LMS–learning management system, SPSS–statistical package for social sciences (software), eigen value, hypothesis
Procedia PDF Downloads 10710026 Sentinel-2 Based Burn Area Severity Assessment Tool in Google Earth Engine
Authors: D. Madhushanka, Y. Liu, H. C. Fernando
Abstract:
Fires are one of the foremost factors of land surface disturbance in diverse ecosystems, causing soil erosion and land-cover changes and atmospheric effects affecting people's lives and properties. Generally, the severity of the fire is calculated as the Normalized Burn Ratio (NBR) index. This is performed manually by comparing two images obtained afterward. Then by using the bitemporal difference of the preprocessed satellite images, the dNBR is calculated. The burnt area is then classified as either unburnt (dNBR<0.1) or burnt (dNBR>= 0.1). Furthermore, Wildfire Severity Assessment (WSA) classifies burnt areas and unburnt areas using classification levels proposed by USGS and comprises seven classes. This procedure generates a burn severity report for the area chosen by the user manually. This study is carried out with the objective of producing an automated tool for the above-mentioned process, namely the World Wildfire Severity Assessment Tool (WWSAT). It is implemented in Google Earth Engine (GEE), which is a free cloud-computing platform for satellite data processing, with several data catalogs at different resolutions (notably Landsat, Sentinel-2, and MODIS) and planetary-scale analysis capabilities. Sentinel-2 MSI is chosen to obtain regular processes related to burnt area severity mapping using a medium spatial resolution sensor (15m). This tool uses machine learning classification techniques to identify burnt areas using NBR and to classify their severity over the user-selected extent and period automatically. Cloud coverage is one of the biggest concerns when fire severity mapping is performed. In WWSAT based on GEE, we present a fully automatic workflow to aggregate cloud-free Sentinel-2 images for both pre-fire and post-fire image compositing. The parallel processing capabilities and preloaded geospatial datasets of GEE facilitated the production of this tool. This tool consists of a Graphical User Interface (GUI) to make it user-friendly. The advantage of this tool is the ability to obtain burn area severity over a large extent and more extended temporal periods. Two case studies were carried out to demonstrate the performance of this tool. The Blue Mountain national park forest affected by the Australian fire season between 2019 and 2020 is used to describe the workflow of the WWSAT. This site detected more than 7809 km2, using Sentinel-2 data, giving an error below 6.5% when compared with the area detected on the field. Furthermore, 86.77% of the detected area was recognized as fully burnt out, of which high severity (17.29%), moderate-high severity (19.63%), moderate-low severity (22.35%), and low severity (27.51%). The Arapaho and Roosevelt National Forest Park, California, the USA, which is affected by the Cameron peak fire in 2020, is chosen for the second case study. It was found that around 983 km2 had burned out, of which high severity (2.73%), moderate-high severity (1.57%), moderate-low severity (1.18%), and low severity (5.45%). These spots also can be detected through the visual inspection made possible by cloud-free images generated by WWSAT. This tool is cost-effective in calculating the burnt area since satellite images are free and the cost of field surveys is avoided.Keywords: burnt area, burnt severity, fires, google earth engine (GEE), sentinel-2
Procedia PDF Downloads 23510025 Improving Radiation Efficiency Using Metamaterial in Pyramidal Horn Antenna
Authors: Amit Kumar Baghel, Sisir Kumar Nayak
Abstract:
The proposed metamaterial design help to increase the radiation efficiency at 2.9 GHz by reducing the side and back lobes by making the phase difference of the waves emerging from the phase center of the horn antenna same after passing through metamaterial array. The unit cell of the metamaterial is having concentric ring structure made of copper of 0.035 mm thickness on both sides of FR4 sheet. The inner ring diameter is kept as 3 mm, and the outer ring diameters are changed according to the path and tramission phase difference of the unit cell from the phase center of the antenna in both the horizontal and vertical direction, i.e., in x- and y-axis. In this case, the ring radius varies from 3.19 mm to 6.99 mm with the respective S21 phase difference of -62.25° to -124.64°. The total phase difference can be calculated by adding the path difference of the respective unit cell in the array to the phase difference of S21. Taking one of the unit cell as the reference, the total phase difference between the reference unit cell and other cells must be integer multiple of 360°. The variation of transmission coefficient S21 with the ring radius is greater than -6 dB. The array having 5 x 5 unit cell is kept inside the pyramidal horn antenna (L X B X H = 295.451 x 384.233 x 298.66 mm3) at a distance of 36.68 mm from the waveguide throat. There is an improvement in side lobe level in E-plane by 14.6 dB when the array is used. The front to back lobe ration is increased by 1 dB by using the array. The proposed antenna with metamaterial array can be used in beam shaping for wireless power transfer applications.Keywords: metamaterial, side lobe level, front to back ratio, beam forming
Procedia PDF Downloads 27410024 Clustering and Modelling Electricity Conductors from 3D Point Clouds in Complex Real-World Environments
Authors: Rahul Paul, Peter Mctaggart, Luke Skinner
Abstract:
Maintaining public safety and network reliability are the core objectives of all electricity distributors globally. For many electricity distributors, managing vegetation clearances from their above ground assets (poles and conductors) is the most important and costly risk mitigation control employed to meet these objectives. Light Detection And Ranging (LiDAR) is widely used by utilities as a cost-effective method to inspect their spatially-distributed assets at scale, often captured using high powered LiDAR scanners attached to fixed wing or rotary aircraft. The resulting 3D point cloud model is used by these utilities to perform engineering grade measurements that guide the prioritisation of vegetation cutting programs. Advances in computer vision and machine-learning approaches are increasingly applied to increase automation and reduce inspection costs and time; however, real-world LiDAR capture variables (e.g., aircraft speed and height) create complexity, noise, and missing data, reducing the effectiveness of these approaches. This paper proposes a method for identifying each conductor from LiDAR data via clustering methods that can precisely reconstruct conductors in complex real-world configurations in the presence of high levels of noise. It proposes 3D catenary models for individual clusters fitted to the captured LiDAR data points using a least square method. An iterative learning process is used to identify potential conductor models between pole pairs. The proposed method identifies the optimum parameters of the catenary function and then fits the LiDAR points to reconstruct the conductors.Keywords: point cloud, LİDAR data, machine learning, computer vision, catenary curve, vegetation management, utility industry
Procedia PDF Downloads 9910023 Cloud Points to Create an Innovative and Custom Ankle Foot Orthosis in CAD Environment
Authors: Y. Benabid, K. Benfriha, V. Rieuf, J. F. Omhover
Abstract:
This paper describes an approach to create custom concepts for innovative products; this approach describes relations between innovation tools and Computer Aided Design environment (use creativity session and design tools). A model for the design process is proposed and explored in order to describe the power tool used to create and ameliorate an innovative product all based upon a range of data (cloud points) in this study. Comparison between traditional method and innovative method we help to generate and put forward a new model of the design process in order to create a custom Ankle Foot Orthosis (AFO) in a CAD environment in order to ameliorate and controlling the motion. The custom concept needs big development in different environments; the relation between these environments is described. The results can help the surgeons in the upstream treatment phases. CAD models can be applied and accepted by professionals in the design and manufacture systems. This development is based on the anatomy of the population of North Africa.Keywords: ankle foot orthosis, CAD, reverse engineering, sketch
Procedia PDF Downloads 45510022 Theoretical Study of Electronic Structure of Erbium (Er), Fermium (Fm), and Nobelium (No)
Authors: Saleh O. Allehabi, V. A. Dzubaa, V. V. Flambaum, Jiguang Li, A. V. Afanasjev, S. E. Agbemava
Abstract:
Recently developed versions of the configuration method for open shells, configuration interaction with perturbation theory (CIPT), and configuration interaction with many-body perturbation theory (CI+MBPT) techniques are used to study the electronic structure of Er, Fm, and No atoms. Excitation energies of odd states connected to the even ground state by electric dipole transitions, the corresponding transition rates, isotope shift, hyperfine structure, ionization potentials, and static scalar polarizabilities are calculated. The way of extracting parameters of nuclear charge distribution beyond nuclear root mean square (RMS) radius, e.g., a parameter of quadrupole deformation β, is demonstrated. In nuclei with spin > 1/2, parameter β is extracted from the quadrupole hyperfine structure. With zero nuclear spin or spin 1/2, it is impossible since quadrupole zero, so a different method was developed. The measurements of at least two atomic transitions are needed to disentangle the contributions of the changes in deformation and nuclear RMS radius into field isotopic shift. This is important for testing nuclear theory and for searching for the hypothetical island of stability. Fm and No are heavy elements approaching the superheavy region, for which the experimental data are very poor, only seven lines for the Fm element and one line for the No element. Since Er and Fm have similar electronic structures, calculations for Er serve as a guide to the accuracy of the calculations. Twenty-eight new levels of Fm atom are reported.Keywords: atomic spectra, electronic transitions, isotope effect, electron correlation calculations for atoms
Procedia PDF Downloads 15510021 Woodcast is Ecologically Sound and Tolerated by a Majority of Patients
Authors: R. Hassan, J. Duncombe, E. Darke, A. Dias, K. Anderson, R. G. Middleton
Abstract:
NHS England has set itself the task of delivering a “Net Zero” National Health service by 2040. It is incumbent upon all health care practioners to work towards this goal. Orthopaedic surgeons are no exception. Distal radial fractures are the most common fractures sustained by the adult population. However, studies are shortcoming on individual patient experience. The aim of this study was to assess the patient’s satisfaction and outcomes with woodcast used in the conservative management of distal radius fractures. For all patients managed with woodcast in our unit, we undertook a structured questionnaire that included the Patient Rated Wrist Evaluation (PRWE) score, The EQ-5D-5L score and the pain numerical score at the time of injury and six weeks after. 30 patients were initially managed with woodcast. 80% of patients tolerated woodcast for the full duration of their treatment. Of these, 20% didn’t tolerate woodcast and had their casts removed within 48 hours. Of the remaining, 79.1% were satisfied about woodcast comfort, 66% were very satisfied about woodcast weight, 70% were satisfied with temperature and sweatiness, 62.5% were very satisfied about the smell/odour, and 75% were satisfied about the level of support woodcast provided. During their treatment, 83.3% of patients rated their pain as five or less. For those who completed their treatment in woodcast, none required any further intervention or utilised the open appointment because of ongoing wrist problems. In conclusion, when woodcast is tolerated, patients’ satisfaction and outcome levels were good. However, we acknowledged 20% of patients in our series were not able to tolerate woodacst, Therefore, we suggest a comparison between the widely used synthetic plaster of Paris casting and woodcast to come in order.Keywords: distal radius fractures, ecological cast, sustainability, woodcast
Procedia PDF Downloads 10110020 Development of R³ UV Exposure for the UV Dose-Insensitive and Cost-Effective Fabrication of Biodegradable Polymer Microneedles
Authors: Sungmin Park, Gyungmok Nam, Seungpyo Woo, Young Choi, Sangheon Park, Sang-Hee Yoon
Abstract:
Puncturing human skin with microneedles is critically important for microneedle-mediate drug delivery. Despite of extensive efforts in the past decades, the scale-up fabrication of sharp-tipped and high-aspect-ratio microneedles, especially made of biodegradable polymers, is still a long way off. Here, we present a UV dose insensitive and cost-effective microfabrication method for the biodegradable polymer microneedles with sharp tips and long lengths which can pierce human skin with low insertion force. The biodegradable polymer microneedles are fabricated with the polymer solution casting where a poly(lactic-co-glycolic acid) (PLGA, 50:50) solution is coated onto a SU-8 mold prepared with a reverse, ramped, and rotational (R3) UV exposure. The R3 UV exposure is modified from the multidirectional UV exposure both to suppress UV reflection from the bottom surface without anti-reflection layers and to optimize solvent concentration in the SU-8 photoresist, therefore achieving robust (i.e., highly insensitive to UV dose) and cost-effective fabrication of biodegradable polymer microneedles. An optical model for describing the spatial distribution of UV irradiation dose of the R3 UV exposure is also developed to theoretically predict the microneedle geometry fabricated with the R3 UV exposure and also to demonstrate the insensitiveness of microneedle geometry to UV dose. In the experimental characterization, the microneedles fabricated with the R3 UV exposure are compared with those fabricated with a conventional method (i.e., multidirectional UV exposure). The R3 UV exposure-based microfabrication reduces the end-tip radius by a factor of 5.8 and the deviation from ideal aspect ratio by 74.8%, compared with conventional method-based microfabrication. The PLGA microneedles fabricated with the R3 UV exposure pierce full-thickness porcine skins successfully and are demonstrated to completely dissolve in PBS (phosphate-buffered saline). The findings of this study will lead to an explosive growth of the microneedle-mediated drug delivery market.Keywords: R³ UV exposure, optical model, UV dose, reflection, solvent concentration, biodegradable polymer microneedle
Procedia PDF Downloads 16610019 Heliport Remote Safeguard System Based on Real-Time Stereovision 3D Reconstruction Algorithm
Authors: Ł. Morawiński, C. Jasiński, M. Jurkiewicz, S. Bou Habib, M. Bondyra
Abstract:
With the development of optics, electronics, and computers, vision systems are increasingly used in various areas of life, science, and industry. Vision systems have a huge number of applications. They can be used in quality control, object detection, data reading, e.g., QR-code, etc. A large part of them is used for measurement purposes. Some of them make it possible to obtain a 3D reconstruction of the tested objects or measurement areas. 3D reconstruction algorithms are mostly based on creating depth maps from data that can be acquired from active or passive methods. Due to the specific appliance in airfield technology, only passive methods are applicable because of other existing systems working on the site, which can be blinded on most spectral levels. Furthermore, reconstruction is required to work long distances ranging from hundreds of meters to tens of kilometers with low loss of accuracy even with harsh conditions such as fog, rain, or snow. In response to those requirements, HRESS (Heliport REmote Safeguard System) was developed; which main part is a rotational head with a two-camera stereovision rig gathering images around the head in 360 degrees along with stereovision 3D reconstruction and point cloud combination. The sub-pixel analysis introduced in the HRESS system makes it possible to obtain an increased distance measurement resolution and accuracy of about 3% for distances over one kilometer. Ultimately, this leads to more accurate and reliable measurement data in the form of a point cloud. Moreover, the program algorithm introduces operations enabling the filtering of erroneously collected data in the point cloud. All activities from the programming, mechanical and optical side are aimed at obtaining the most accurate 3D reconstruction of the environment in the measurement area.Keywords: airfield monitoring, artificial intelligence, stereovision, 3D reconstruction
Procedia PDF Downloads 12410018 Developing a Framework for Open Source Software Adoption in a Higher Education Institution in Uganda. A case of Kyambogo University
Authors: Kafeero Frank
Abstract:
This study aimed at developing a frame work for open source software adoption in an institution of higher learning in Uganda, with the case of KIU as a study area. There were mainly four research questions based on; individual staff interaction with open source software forum, perceived FOSS characteristics, organizational characteristics and external characteristics as factors that affect open source software adoption. The researcher used causal-correlation research design to study effects of these variables on open source software adoption. A quantitative approach was used in this study with self-administered questionnaire on a purposively and randomly sampled sample of university ICT staff. Resultant data was analyzed using means, correlation coefficients and multivariate multiple regression analysis as statistical tools. The study reveals that individual staff interaction with open source software forum and perceived FOSS characteristics were the primary factors that significantly affect FOSS adoption while organizational and external factors were secondary with no significant effect but significant correlation to open source software adoption. It was concluded that for effective open source software adoption to occur there must be more effort on primary factors with subsequent reinforcement of secondary factors to fulfill the primary factors and adoption of open source software. Lastly recommendations were made in line with conclusions for coming up with Kyambogo University frame work for open source software adoption in institutions of higher learning. Areas of further research recommended include; Stakeholders’ analysis of open source software adoption in Uganda; Challenges and way forward. Evaluation of Kyambogo University frame work for open source software adoption in institutions of higher learning. Framework development for cloud computing adoption in Ugandan universities. Framework for FOSS development in Uganda IT industryKeywords: open source software., organisational characteristics, external characteristics, cloud computing adoption
Procedia PDF Downloads 7210017 A Simple Technique for Centralisation of Distal Femoral Nail to Avoid Anterior Femoral Impingement and Perforation
Authors: P. Panwalkar, K. Veravalli, M. Tofighi, A. Mofidi
Abstract:
Introduction: Anterior femoral perforation or distal anterior nail position is a known complication of femoral nailing specifically in pertrochantric fractures fixed with cephalomedullary nail. This has been attributed to wrong entry point for the femoral nail, nail with large radius of curvature or malreduced fracture. Left alone anterior perforation of femur or abutment of nail on anterior femur will result in pain and risk stress riser at distal femur and periprosthetic fracture. There have been multiple techniques described to avert or correct this problem ranging from using different nail, entry point change, poller screw to deflect the nail position, use of shorter nail or use of curved guidewire or change of nail to ensure a nail with large radius of curvature Methods: We present this technique which we have used in order to centralise the femoral nail either when the nail has been put anteriorly or when the guide wire has been inserted too anteriorly prior to the insertion of the nail. This technique requires the use of femoral reduction spool from the nailing set. This technique was used by eight trainees of different level of experience under supervision. Results: This technique was easily reproducible without any learning curve without a need for opening of fracture site or change in the entry point with three different femoral nailing sets in twenty-five cases. The process took less than 10 minutes even when revising a malpositioned femoral nail. Conclusion: Our technique of using femoral reduction spool is easily reproducible and repeatable technique for avoidance of non-centralised femoral nail insertion and distal anterior perforation of femoral nail.Keywords: femoral fracture, nailing, malposition, surgery
Procedia PDF Downloads 14210016 MLOps Scaling Machine Learning Lifecycle in an Industrial Setting
Authors: Yizhen Zhao, Adam S. Z. Belloum, Goncalo Maia Da Costa, Zhiming Zhao
Abstract:
Machine learning has evolved from an area of academic research to a real-word applied field. This change comes with challenges, gaps and differences exist between common practices in academic environments and the ones in production environments. Following continuous integration, development and delivery practices in software engineering, similar trends have happened in machine learning (ML) systems, called MLOps. In this paper we propose a framework that helps to streamline and introduce best practices that facilitate the ML lifecycle in an industrial setting. This framework can be used as a template that can be customized to implement various machine learning experiment. The proposed framework is modular and can be recomposed to be adapted to various use cases (e.g. data versioning, remote training on cloud). The framework inherits practices from DevOps and introduces other practices that are unique to the machine learning system (e.g.data versioning). Our MLOps practices automate the entire machine learning lifecycle, bridge the gap between development and operation.Keywords: cloud computing, continuous development, data versioning, DevOps, industrial setting, MLOps
Procedia PDF Downloads 26510015 Relationship between Effective Classroom Management with Students’ Academic Achievement of EFL of STKIP YPUP
Authors: Eny Syatriana
Abstract:
The purpose of this study is to find out the effective instruction for classroom management, with the main identification of organizing and managing effective learning environments, to identify characteristics of effective lesson planning, identify resources and materials dealing with positive and effective classroom management. Knowing the effective instruction management is one of the characteristics of well managed teacher. The study was carried out in three randomly selected classes of STKIP YPUP in South Sulawesi. The design adopted for the study was a descriptive survey approach. Simple descriptive analysis was used. The major instrument used in this study were student questionnaire, teacher questionnaire, data were gathered with the research instrument and were analyzed, the research question were investigated and two hypothesis were duly tested using t-test statistics. Based on the findings of this research, it was concluded that effective classroom management skills or techniques have strong and positive influence on student achievement.Keywords: effective classroom management skills, students’ achievement, students academic, effective learning environments
Procedia PDF Downloads 33310014 Advanced Model for Calculation of the Neutral Axis Shifting and the Wall Thickness Distribution in Rotary Draw Bending Processes
Abstract:
Rotary draw bending is a method which is being used in tube forming. In the tube bending process, the neutral axis moves towards the inner arc and the wall thickness distribution changes for tube’s cross section. Thinning takes place in the outer arc of the tube (extrados) due to the stretching of the material, whereas thickening occurs in the inner arc of the tube (intrados) due to the comparison of the material. The calculations of the wall thickness distribution, neutral axis shifting, and strain distribution have not been accurate enough, so far. The previous model (the geometrical model) describes the neutral axis shifting and wall thickness distribution. The geometrical of the tube, bending radius and bending angle are considered in the geometrical model, while the influence of the material properties of the tube forming are ignored. The advanced model is a modification of the previous model using material properties that depends on the correction factor. The correction factor is a purely empirically determined factor. The advanced model was compared with the Finite element simulation (FE simulation) using a different bending factor (Bf=bending radius/ diameter of the tube), wall thickness (Wf=diameter of the tube/ wall thickness), and material properties (strain hardening exponent). Finite element model of rotary draw bending has been performed in PAM-TUBE program (version: 2012). Results from the advanced model resemble the FE simulation and the experimental test.Keywords: rotary draw bending, material properties, neutral axis shifting, wall thickness distribution
Procedia PDF Downloads 39710013 Woodcast Is Ecologically Sound and Tolerated by Majority of Patients
Authors: R. Hassan, J. Duncombe, E. Darke, A. Dias, K. Anderson, R. G. Middleton
Abstract:
Background: NHS England has set itself the task of delivering a “Net Zero” National Health service by 2040. It is incumbent upon all health care practioners to work towards this goal. Orthopaedic surgeons are no exception. Distal radial fractures are the most common fractures sustained by the adult population. However, studiesare shortcoming on individual patient experience. The aim of this study was to assess the patient’ssatisfaction and outcomes with woodcast used in the conservative management of distal radius fractures. Methods: For all patients managed with woodcast in our unit, we undertook a structured questionnairethat included the Patient Rated Wrist Evaluation (PRWE) score, The EQ-5D-5L score, and the pain numerical score at the time of injury and six weeks after. Results: 30 patients were initially managed with woodcast.80% of patients tolerated woodcast for the full duration of their treatment. Of these, 20% didn’t tolerate woodcast and had their casts removed within 48 hours. Of the remaining, 79.1% were satisfied about woodcast comfort, 66% were very satisfied about woodcast weight, 70% were satisfied with temperature and sweatiness, 62.5% were very satisfied about the smell/odour, and 75% were satisfied about the level of support woodcast provided. During their treatment, 83.3% of patients rated their pain as five or less. Conclusion: For those who completed their treatment in woodcast, none required any further intervention or utilised the open appointment because of ongoing wrist problems. In conclusion, when woodcast is tolerated, patients’ satisfaction and outcome levels were good. However, we acknowledged 20% of patients in our series were not able to tolerate woodacst, Therefore, we suggest a comparison between the widely used synthetic plaster of Pariscasting and woodcast to come in order.Keywords: distal radius fractures, ecological cast, sustainability, woodcast
Procedia PDF Downloads 8510012 Learning to Teach on the Cloud: Preservice EFL Teachers’ Online Project-Based Practicum Experience
Authors: Mei-Hui Liu
Abstract:
This paper reports 20 preservice EFL teachers’ learning-to-teach experience when they were engaged in an online project-based practicum implemented on a Cloud Platform. This 10-month study filled in the literature gap by documenting the impact of online project-based instruction on preservice EFL teachers’ professional development. Data analysis showed that the online practicum was regarded as a flexible mechanism offering chances of teaching practices without geographical barriers. Additionally, this project-based practice helped the participants integrate the theories they had learned and further foster them how to create a self-directed online learning environment. Furthermore, these preservice teachers with experiences of technology-enabled practicum showed their motivation to apply technology and online platforms into future instructional practices. Yet, this study uncovered several concerns encountered by these participants during this online field experience. The findings of this study rendered meaning and lessons for teacher educators intending to integrate online practicum into preservice training courses.Keywords: online teaching practicum, project-based learning, teacher preparation, English language education
Procedia PDF Downloads 37110011 Approaches to Ethical Hacking: A Conceptual Framework for Research
Authors: Lauren Provost
Abstract:
The digital world remains increasingly vulnerable, making the development of effective cybersecurity approaches even more critical in supporting the success of the digital economy and national security. Although approaches to cybersecurity have shifted and improved in the last decade with new models, especially with cloud computing and mobility, a record number of high severity vulnerabilities were recorded in the National Institute of Standards and Technology (NIST), and its National Vulnerability Database (NVD) in 2020. This is due, in part, to the increasing complexity of cyber ecosystems. Security must be approached with a more comprehensive, multi-tool strategy that addresses the complexity of cyber ecosystems, including the human factor. Ethical hacking has emerged as such an approach: a more effective, multi-strategy, comprehensive approach to cyber security's most pressing needs, especially understanding the human factor. Research on ethical hacking, however, is limited in scope. The two main objectives of this work are to (1) provide highlights of case studies in ethical hacking, (2) provide a conceptual framework for research in ethical hacking that embraces and addresses both technical and nontechnical security measures. Recommendations include an improved conceptual framework for research centered on ethical hacking that addresses many factors and attributes of significant attacks that threaten computer security; a more robust, integrative multi-layered framework embracing the complexity of cybersecurity ecosystems.Keywords: ethical hacking, literature review, penetration testing, social engineering
Procedia PDF Downloads 21810010 A Cloud-Based Federated Identity Management in Europe
Authors: Jesus Carretero, Mario Vasile, Guillermo Izquierdo, Javier Garcia-Blas
Abstract:
Currently, there is a so called ‘identity crisis’ in cybersecurity caused by the substantial security, privacy and usability shortcomings encountered in existing systems for identity management. Federated Identity Management (FIM) could be solution for this crisis, as it is a method that facilitates management of identity processes and policies among collaborating entities without enforcing a global consistency, that is difficult to achieve when there are ID legacy systems. To cope with this problem, the Connecting Europe Facility (CEF) initiative proposed in 2014 a federated solution in anticipation of the adoption of the Regulation (EU) N°910/2014, the so-called eIDAS Regulation. At present, a network of eIDAS Nodes is being deployed at European level to allow that every citizen recognized by a member state is to be recognized within the trust network at European level, enabling the consumption of services in other member states that, until now were not allowed, or whose concession was tedious. This is a very ambitious approach, since it tends to enable cross-border authentication of Member States citizens without the need to unify the authentication method (eID Scheme) of the member state in question. However, this federation is currently managed by member states and it is initially applied only to citizens and public organizations. The goal of this paper is to present the results of a European Project, named eID@Cloud, that focuses on the integration of eID in 5 cloud platforms belonging to authentication service providers of different EU Member States to act as Service Providers (SP) for private entities. We propose an initiative based on a private eID Scheme both for natural and legal persons. The methodology followed in the eID@Cloud project is that each Identity Provider (IdP) is subscribed to an eIDAS Node Connector, requesting for authentication, that is subscribed to an eIDAS Node Proxy Service, issuing authentication assertions. To cope with high loads, load balancing is supported in the eIDAS Node. The eID@Cloud project is still going on, but we already have some important outcomes. First, we have deployed the federation identity nodes and tested it from the security and performance point of view. The pilot prototype has shown the feasibility of deploying this kind of systems, ensuring good performance due to the replication of the eIDAS nodes and the load balance mechanism. Second, our solution avoids the propagation of identity data out of the native domain of the user or entity being identified, which avoids problems well known in cybersecurity due to network interception, man in the middle attack, etc. Last, but not least, this system allows to connect any country or collectivity easily, providing incremental development of the network and avoiding difficult political negotiations to agree on a single authentication format (which would be a major stopper).Keywords: cybersecurity, identity federation, trust, user authentication
Procedia PDF Downloads 16610009 Terrestrial Laser Scans to Assess Aerial LiDAR Data
Authors: J. F. Reinoso-Gordo, F. J. Ariza-López, A. Mozas-Calvache, J. L. García-Balboa, S. Eddargani
Abstract:
The DEMs quality may depend on several factors such as data source, capture method, processing type used to derive them, or the cell size of the DEM. The two most important capture methods to produce regional-sized DEMs are photogrammetry and LiDAR; DEMs covering entire countries have been obtained with these methods. The quality of these DEMs has traditionally been evaluated by the national cartographic agencies through punctual sampling that focused on its vertical component. For this type of evaluation there are standards such as NMAS and ASPRS Positional Accuracy Standards for Digital Geospatial Data. However, it seems more appropriate to carry out this evaluation by means of a method that takes into account the superficial nature of the DEM and, therefore, its sampling is superficial and not punctual. This work is part of the Research Project "Functional Quality of Digital Elevation Models in Engineering" where it is necessary to control the quality of a DEM whose data source is an experimental LiDAR flight with a density of 14 points per square meter to which we call Point Cloud Product (PCpro). In the present work it is described the capture data on the ground and the postprocessing tasks until getting the point cloud that will be used as reference (PCref) to evaluate the PCpro quality. Each PCref consists of a patch 50x50 m size coming from a registration of 4 different scan stations. The area studied was the Spanish region of Navarra that covers an area of 10,391 km2; 30 patches homogeneously distributed were necessary to sample the entire surface. The patches have been captured using a Leica BLK360 terrestrial laser scanner mounted on a pole that reached heights of up to 7 meters; the position of the scanner was inverted so that the characteristic shadow circle does not exist when the scanner is in direct position. To ensure that the accuracy of the PCref is greater than that of the PCpro, the georeferencing of the PCref has been carried out with real-time GNSS, and its accuracy positioning was better than 4 cm; this accuracy is much better than the altimetric mean square error estimated for the PCpro (<15 cm); The kind of DEM of interest is the corresponding to the bare earth, so that it was necessary to apply a filter to eliminate vegetation and auxiliary elements such as poles, tripods, etc. After the postprocessing tasks the PCref is ready to be compared with the PCpro using different techniques: cloud to cloud or after a resampling process DEM to DEM.Keywords: data quality, DEM, LiDAR, terrestrial laser scanner, accuracy
Procedia PDF Downloads 10010008 The Impact of Direct and Indirect Pressure Measuring Systems on the Pressure Mapping for the Medical Compression Garments
Authors: Arash M. Shahidi, Tilak Dias, Gayani K. Nandasiri
Abstract:
While graduated compression is the foundation of treatment and management of many medical complications such as leg ulcer, varicose veins, and lymphedema, monitoring the interface pressure has been conducted using different sensors that operate based on diverse approaches. The variations existed from the pressure readings collected using different interface pressure measurement systems would cause difficulties in taking a decision regarding the compression therapy. It is crucial to acknowledge the differences existing between direct and indirect pressure measurement systems while considering the commercially available systems such as AMI, Picopress and OPM which are under direct measurements systems, and HATRA (BSI), HOSY (RAL-GZ) and FlexiForce which comes under the indirect measurement system. Furthermore, Piezo-resistive sensors (Flexiforce) can measure the changes in resistance corresponding to the applied force on the sensing area. Direct pressure measuring systems are capable of measuring interface pressure on the three-dimensional states, while the indirect pressure measuring systems stretch the fabric in the two-dimensional direction and extrapolate pressure from surface tension measured on the device and neglect the vital factor which is the radius of curvature. In this study, a leg mannequin of known dimensions is selected with a knitted class 3 compression stocking. It has been decided to evaluate the data collected from different available systems (AMI, PicoPress, FlexiForce, and HATRA) and compare the results. The results showed a discrepancy between Hatra, AMI, Picopress, and Flexiforce against the pressure standard used to generate class 3 compression stocking. As predicted a higher pressure value with direct interface measuring systems were monitored against HATRA due to the effect of the radius of curvature.Keywords: AMI, FlexiForce, graduated compression, HATRA, interface pressure, PicoPress
Procedia PDF Downloads 35210007 Designing a Dispersion Flattened Single Mode PCF for E-Band to U-Band with Less Effective Area
Authors: Shabbir Chowdhury
Abstract:
A signal is broadened when it is gone through a channel, this phenomenon is known as dispersion. And dispersion is different for different wavelength. So bandwidth become limited. Research have tried to design an optical fiber with flattened dispersion to use more bandwidth and also for wavelength division multiplexing. In this paper, a single mode photonic crystal fiber with a flattened dispersion and less effective area has been proposed where silica is used as fiber materials. The effective dispersion varies from -1.996 to 0.1783 [ps/(nm-km)] for enter E-band to U-band. This fiber will take only 3.048 [micrometer^2] (for 1.75 micrometer wavelength). Silica is being used as the fiber material.Keywords: photonic crystal fiber, dispersion, bandwidth, chromatic dispersion, effective dispersion, dispersion compensation, effective area, effective refractive index
Procedia PDF Downloads 41510006 Structural Behavior of Composite Hollow RC Column under Combined Loads
Authors: Abdul Qader Melhm, Hussein Elrafidi
Abstract:
This paper is dealing with studying the structural behavior of a steel-composite hollow reinforced concrete (RC) column model under combined eccentric loading. The composite model consists of an inner steel tube surrounded via a concrete core with longitudinal and circular transverse reinforcement. The radius of gyration according to American and Euro specifications be calculated, in order to calculate the thinnest ratio for this type of composite column model, in addition to the flexural rigidity. Formulas for interaction diagram is given for this type of model, which is a general loading conditions in which an element is exposed to an axial load with bending at the same time. The structural capacity of this model, elastic, plastic loads and strains will be computed and compared with experimental results. The total eccentric axial load of the column model is calculated based on the effective length KL available from several relationships provided in the paper. Furthermore, the inner tube experiences buckling failure after reaching its maximum strength will be investigated.Keywords: column, composite, eccentric, inner tube, interaction, reinforcement
Procedia PDF Downloads 19210005 Cyber Attacks Management in IoT Networks Using Deep Learning and Edge Computing
Authors: Asmaa El Harat, Toumi Hicham, Youssef Baddi
Abstract:
This survey delves into the complex realm of Internet of Things (IoT) security, highlighting the urgent need for effective cybersecurity measures as IoT devices become increasingly common. It explores a wide array of cyber threats targeting IoT devices and focuses on mitigating these attacks through the combined use of deep learning and machine learning algorithms, as well as edge and cloud computing paradigms. The survey starts with an overview of the IoT landscape and the various types of attacks that IoT devices face. It then reviews key machine learning and deep learning algorithms employed in IoT cybersecurity, providing a detailed comparison to assist in selecting the most suitable algorithms. Finally, the survey provides valuable insights for cybersecurity professionals and researchers aiming to enhance security in the intricate world of IoT.Keywords: internet of things (IoT), cybersecurity, machine learning, deep learning
Procedia PDF Downloads 3110004 An Exploratory Study Applied to Search Relationship between Humans and Universe
Authors: Mohamed Hashelaf, Ahmed Al-Osdody
Abstract:
In this paper, we focused our efforts on one of the vaguest subjects in astrophysics that is the formation and evolution of the universe until the arrival of humans. Through an in-depth exploration of the origins of the universe, understanding what has happened since the Big Bang until now and checking the history of creation, we can answer questions about the future of life, the possibility of its existence elsewhere in the universe and to be able to understand how we came, what our role in the circle of life is and what the future of our development will be. Here is where we used systematic steps that allowed us first and foremost to identify the reason behind the big bang itself that formed a large cloud of cosmic dust. Then after a period of time from the expansion of the universe and its coolness, the initial molecules of gases from the cosmic cloud began to condense, forming a very dense field of gravity that after millions of years led to the formation of stars, galaxies, even earth and the else planets. Finally, it became clear before us that after the earth has formed, the existence of liquid water made it possible for life to form, starting from the bacteria all the way until the appearance of the humans that we know today. But it does not stop here. If we look and contemplate in ourselves as humans, we will understand that the universe is inside us and that’s what makes us exceptional. All of this means that just as life on earth was created, it could have been on other planets as well. It also means that we are the universe’s key to understand itself.Keywords: Big Bang, cosmic dust, primary elements, universe
Procedia PDF Downloads 13410003 Automated Transformation of 3D Point Cloud to BIM Model: Leveraging Algorithmic Modeling for Efficient Reconstruction
Authors: Radul Shishkov, Orlin Davchev
Abstract:
The digital era has revolutionized architectural practices, with building information modeling (BIM) emerging as a pivotal tool for architects, engineers, and construction professionals. However, the transition from traditional methods to BIM-centric approaches poses significant challenges, particularly in the context of existing structures. This research introduces a technical approach to bridge this gap through the development of algorithms that facilitate the automated transformation of 3D point cloud data into detailed BIM models. The core of this research lies in the application of algorithmic modeling and computational design methods to interpret and reconstruct point cloud data -a collection of data points in space, typically produced by 3D scanners- into comprehensive BIM models. This process involves complex stages of data cleaning, feature extraction, and geometric reconstruction, which are traditionally time-consuming and prone to human error. By automating these stages, our approach significantly enhances the efficiency and accuracy of creating BIM models for existing buildings. The proposed algorithms are designed to identify key architectural elements within point clouds, such as walls, windows, doors, and other structural components, and to translate these elements into their corresponding BIM representations. This includes the integration of parametric modeling techniques to ensure that the generated BIM models are not only geometrically accurate but also embedded with essential architectural and structural information. Our methodology has been tested on several real-world case studies, demonstrating its capability to handle diverse architectural styles and complexities. The results showcase a substantial reduction in time and resources required for BIM model generation while maintaining high levels of accuracy and detail. This research contributes significantly to the field of architectural technology by providing a scalable and efficient solution for the integration of existing structures into the BIM framework. It paves the way for more seamless and integrated workflows in renovation and heritage conservation projects, where the accuracy of existing conditions plays a critical role. The implications of this study extend beyond architectural practices, offering potential benefits in urban planning, facility management, and historic preservation.Keywords: BIM, 3D point cloud, algorithmic modeling, computational design, architectural reconstruction
Procedia PDF Downloads 6310002 Nature of Forest Fragmentation Owing to Human Population along Elevation Gradient in Different Countries in Hindu Kush Himalaya Mountains
Authors: Pulakesh Das, Mukunda Dev Behera, Manchiraju Sri Ramachandra Murthy
Abstract:
Large numbers of people living in and around the Hindu Kush Himalaya (HKH) region, depends on this diverse mountainous region for ecosystem services. Following the global trend, this region also experiencing rapid population growth, and demand for timber and agriculture land. The eight countries sharing the HKH region have different forest resources utilization and conservation policies that exert varying forces in the forest ecosystem. This created a variable spatial as well altitudinal gradient in rate of deforestation and corresponding forest patch fragmentation. The quantitative relationship between fragmentation and demography has not been established before for HKH vis-à-vis along elevation gradient. This current study was carried out to attribute the overall and different nature in landscape fragmentations along the altitudinal gradient with the demography of each sharing countries. We have used the tree canopy cover data derived from Landsat data to analyze the deforestation and afforestation rate, and corresponding landscape fragmentation observed during 2000 – 2010. Area-weighted mean radius of gyration (AMN radius of gyration) was computed owing to its advantage as spatial indicator of fragmentation over non-spatial fragmentation indices. Using the subtraction method, the change in fragmentation was computed during 2000 – 2010. Using the tree canopy cover data as a surrogate of forest cover, highest forest loss was observed in Myanmar followed by China, India, Bangladesh, Nepal, Pakistan, Bhutan, and Afghanistan. However, the sequence of fragmentation was different after the maximum fragmentation observed in Myanmar followed by India, China, Bangladesh, and Bhutan; whereas increase in fragmentation was seen following the sequence of as Nepal, Pakistan, and Afghanistan. Using SRTM-derived DEM, we observed higher rate of fragmentation up to 2400m that corroborated with high human population for the year 2000 and 2010. To derive the nature of fragmentation along the altitudinal gradients, the Statistica software was used, where the user defined function was utilized for regression applying the Gauss-Newton estimation method with 50 iterations. We observed overall logarithmic decrease in fragmentation change (area-weighted mean radius of gyration), forest cover loss and population growth during 2000-2010 along the elevation gradient with very high R2 values (i.e., 0.889, 0.895, 0.944 respectively). The observed negative logarithmic function with the major contribution in the initial elevation gradients suggest to gap filling afforestation in the lower altitudes to enhance the forest patch connectivity. Our finding on the pattern of forest fragmentation and human population across the elevation gradient in HKH region will have policy level implication for different nations and would help in characterizing hotspots of change. Availability of free satellite derived data products on forest cover and DEM, grid-data on demography, and utility of geospatial tools helped in quick evaluation of the forest fragmentation vis-a-vis human impact pattern along the elevation gradient in HKH.Keywords: area-weighted mean radius of gyration, fragmentation, human impact, tree canopy cover
Procedia PDF Downloads 21510001 Large Eddy Simulation of Particle Clouds Using Open-Source CFD
Authors: Ruo-Qian Wang
Abstract:
Open-source CFD has become increasingly popular and promising. The recent progress in multiphase flow enables new CFD applications, which provides an economic and flexible research tool for complex flow problems. Our numerical study using four-way coupling Euler-Lagrangian Large-Eddy Simulations to resolve particle cloud dynamics with OpenFOAM and CFDEM will be introduced: The fractioned Navier-Stokes equations are numerically solved for fluid phase motion, solid phase motion is addressed by Lagrangian tracking for every single particle, and total momentum is conserved by fluid-solid inter-phase coupling. The grid convergence test was performed, which proves the current resolution of the mesh is appropriate. Then, we validated the code by comparing numerical results with experiments in terms of particle cloud settlement and growth. A good comparison was obtained showing reliability of the present numerical schemes. The time and height at phase separations were defined and analyzed for a variety of initial release conditions. Empirical formulas were drawn to fit the results.Keywords: four-way coupling, dredging, land reclamation, multiphase flows, oil spill
Procedia PDF Downloads 42910000 Particle Observation in Secondary School Using a Student-Built Instrument: Design-Based Research on a STEM Sequence about Particle Physics
Authors: J.Pozuelo-Muñoz, E. Cascarosa-Salillas, C. Rodríguez-Casals, A. de Echave, E. Terrado-Sieso
Abstract:
This study focuses on the development, implementation, and evaluation of an instructional sequence aimed at 16–17-year-old students, involving the design and use of a cloud chamber—a device that allows observation of subatomic particles. The research addresses the limited presence of particle physics in Spanish secondary and high school curricula, a gap that restricts students' learning of advanced physics concepts and diminishes engagement with complex scientific topics. The primary goal of this project is to introduce particle physics in the classroom through a practical, interdisciplinary methodology that promotes autonomous learning and critical thinking. The methodology is framed within Design-Based Research (DBR), an approach that enables iterative and pragmatic development of educational resources. The research proceeded in several phases, beginning with the design of an experimental teaching sequence, followed by its implementation in high school classrooms. This sequence was evaluated, redesigned, and reimplemented with the aim of enhancing students’ understanding and skills related to designing and using particle detection instruments. The instructional sequence was divided into four stages: introduction to the activity, research and design of cloud chamber prototypes, observation of particle tracks, and analysis of collected data. In the initial stage, students were introduced to the fundamentals of the activity and provided with bibliographic resources to conduct autonomous research on cloud chamber functioning principles. During the design stage, students sourced materials and constructed their own prototypes, stimulating creativity and understanding of physics concepts like thermodynamics and material properties. The third stage focused on observing subatomic particles, where students recorded and analyzed the tracks generated in their chambers. Finally, critical reflection was encouraged regarding the instrument's operation and the nature of the particles observed. The results show that designing the cloud chamber motivates students and actively engages them in the learning process. Additionally, the use of this device introduces advanced scientific topics beyond particle physics, promoting a broader understanding of science. The study’s conclusions emphasize the need to provide students with ample time and space to thoroughly understand the role of materials and physical conditions in the functioning of their prototypes and to encourage critical analysis of the obtained data. This project not only highlights the importance of interdisciplinarity in science education but also provides a practical framework for teachers to adapt complex concepts for educational contexts where these topics are often absent.Keywords: cloud chamber, particle physics, secondary education, instructional design, design-based research, STEM
Procedia PDF Downloads 13