Search results for: analysis and real time information about liquefaction
45164 Real Estate Trend Prediction with Artificial Intelligence Techniques
Authors: Sophia Liang Zhou
Abstract:
For investors, businesses, consumers, and governments, an accurate assessment of future housing prices is crucial to critical decisions in resource allocation, policy formation, and investment strategies. Previous studies are contradictory about macroeconomic determinants of housing price and largely focused on one or two areas using point prediction. This study aims to develop data-driven models to accurately predict future housing market trends in different markets. This work studied five different metropolitan areas representing different market trends and compared three-time lagging situations: no lag, 6-month lag, and 12-month lag. Linear regression (LR), random forest (RF), and artificial neural network (ANN) were employed to model the real estate price using datasets with S&P/Case-Shiller home price index and 12 demographic and macroeconomic features, such as gross domestic product (GDP), resident population, personal income, etc. in five metropolitan areas: Boston, Dallas, New York, Chicago, and San Francisco. The data from March 2005 to December 2018 were collected from the Federal Reserve Bank, FBI, and Freddie Mac. In the original data, some factors are monthly, some quarterly, and some yearly. Thus, two methods to compensate missing values, backfill or interpolation, were compared. The models were evaluated by accuracy, mean absolute error, and root mean square error. The LR and ANN models outperformed the RF model due to RF’s inherent limitations. Both ANN and LR methods generated predictive models with high accuracy ( > 95%). It was found that personal income, GDP, population, and measures of debt consistently appeared as the most important factors. It also showed that technique to compensate missing values in the dataset and implementation of time lag can have a significant influence on the model performance and require further investigation. The best performing models varied for each area, but the backfilled 12-month lag LR models and the interpolated no lag ANN models showed the best stable performance overall, with accuracies > 95% for each city. This study reveals the influence of input variables in different markets. It also provides evidence to support future studies to identify the optimal time lag and data imputing methods for establishing accurate predictive models.Keywords: linear regression, random forest, artificial neural network, real estate price prediction
Procedia PDF Downloads 10345163 Object Recognition System Operating from Different Type Vehicles Using Raspberry and OpenCV
Authors: Maria Pavlova
Abstract:
In our days, it is possible to put the camera on different vehicles like quadcopter, train, airplane and etc. The camera also can be the input sensor in many different systems. That means the object recognition like non separate part of monitoring control can be key part of the most intelligent systems. The aim of this paper is to focus of the object recognition process during vehicles movement. During the vehicle’s movement the camera takes pictures from the environment without storage in Data Base. In case the camera detects a special object (for example human or animal), the system saves the picture and sends it to the work station in real time. This functionality will be very useful in emergency or security situations where is necessary to find a specific object. In another application, the camera can be mounted on crossroad where do not have many people and if one or more persons come on the road, the traffic lights became the green and they can cross the road. In this papers is presented the system has solved the aforementioned problems. It is presented architecture of the object recognition system includes the camera, Raspberry platform, GPS system, neural network, software and Data Base. The camera in the system takes the pictures. The object recognition is done in real time using the OpenCV library and Raspberry microcontroller. An additional feature of this library is the ability to display the GPS coordinates of the captured objects position. The results from this processes will be sent to remote station. So, in this case, we can know the location of the specific object. By neural network, we can learn the module to solve the problems using incoming data and to be part in bigger intelligent system. The present paper focuses on the design and integration of the image recognition like a part of smart systems.Keywords: camera, object recognition, OpenCV, Raspberry
Procedia PDF Downloads 21845162 Numerical Modeling of the Seismic Site Response in the Firenze Metropolitan Area
Authors: Najmeh Ayoqi, Emanuele Marchetti
Abstract:
OpenSWPC was used to model 2D and 3D seismic waveforms produced by various earthquakes in the Firenze metropolitan area. OpenSWPC is an Opens source code for simulation of seismic wave by using the finite difference method (FDM) in Message Passing Interface (MPI) environment. it considered both earthquake sources, with variable magnitude and location, as well as a pulse source in the modeling domain, which is optimal to simulate local seismic amplification effects. Multiple tests were performed to evaluate the dependence of the frequency content of output modeled waveforms on the model grid size and time steps . Moreover the effect of the velocity structure and absorbing boundary condition on waveform features (amplitude, duration and frequency content) where analysed. Eventually model results are compared with real waveform and Horizontal-to-Vertical spectral Ratio (HVSR) , showing that seismic wave modeling can provide important information on seismic assessment in the city.Keywords: openSWPC, earthquake, firenze, HVSR, seismic wave
Procedia PDF Downloads 1845161 Analyzing the Effectiveness of a Bank of Parallel Resistors, as a Burden Compensation Technique for Current Transformer's Burden, Using LabVIEW™ Data Acquisition Tool
Authors: Dilson Subedi
Abstract:
Current transformers are an integral part of power system because it provides a proportional safe amount of current for protection and measurement applications. However, due to upgradation of electromechanical relays to numerical relays and electromechanical energy meters to digital meters, the connected burden, which defines some of the CT characteristics, has drastically reduced. This has led to the system experiencing high currents damaging the connected relays and meters. Since the protection and metering equipment's are designed to withstand only certain amount of current with respect to time, these high currents pose a risk to man and equipment. Therefore, during such instances, the CT saturation characteristics have a huge influence on the safety of both man and equipment and on the reliability of the protection and metering system. This paper shows the effectiveness of a bank of parallel connected resistors, as a burden compensation technique, in compensating the burden of under-burdened CT’s. The response of the CT in the case of failure of one or more resistors at different levels of overcurrent will be captured using the LabVIEWTM data acquisition hardware (DAQ). The analysis is done on the real-time data gathered using LabVIEWTM. Variation of current transformer saturation characteristics with changes in burden will be discussed.Keywords: accuracy limiting factor, burden, burden compensation, current transformer
Procedia PDF Downloads 24545160 The Impacts of Digital Marketing Activities on Customers' Purchase Intention via Brand Reputation and Awareness: Empirical Study
Authors: Radwan Al Dwairi, Sara Melhem
Abstract:
Today’s billions of individuals are linked together in real-time using different types of social platforms. Despite the increasing importance of social media marketing activities in enhancing customers’ intention to purchase online; still, the majority of research has concentrated on the impact of such tools on customer satisfaction or retention and neglecting its real role in enhancing brand reputation and awareness, which in turn impact customers’ intention to purchase online. In response, this study aims to close this gap by conducting an empirical study using a qualitative approach by collecting a sample of data from 216 respondents in this domain. Results of the study reveal the significant impact of word-of-mouth, interactions, and influencers on a brand reputation, where the latter positively and significantly impacted customers’ intention to purchase via social platforms. In addition, results show the significant impact of brand reputation on enhancing customers' purchase intention.Keywords: brand awareness, brand reputation, EWOM, influencers, interaction
Procedia PDF Downloads 9645159 Power Flow and Modal Analysis of a Power System Including Unified Power Flow Controller
Authors: Djilani Kobibi Youcef Islam, Hadjeri Samir, Djehaf Mohamed Abdeldjalil
Abstract:
The Flexible AC Transmission System (FACTS) technology is a new advanced solution that increases the reliability and provides more flexibility, controllability, and stability of a power system. The Unified Power Flow Controller (UPFC), as the most versatile FACTS device for regulating power flow, is able to control respectively transmission line real power, reactive power, and node voltage. The main purpose of this paper is to analyze the effect of the UPFC on the load flow, the power losses, and the voltage stability using NEPLAN software modules, Newton-Raphson load flow is used for the power flow analysis and the modal analysis is used for the study of the voltage stability. The simulation was carried out on the IEEE 14-bus test system.Keywords: FACTS, load flow, modal analysis, UPFC, voltage stability
Procedia PDF Downloads 51645158 Drive Sharing with Multimodal Interaction: Enhancing Safety and Efficiency
Authors: Sagar Jitendra Mahendrakar
Abstract:
Exploratory testing is a dynamic and adaptable method of software quality assurance that is frequently praised for its ability to find hidden flaws and improve the overall quality of the product. Instead of using preset test cases, exploratory testing allows testers to explore the software application dynamically. This is in contrast to scripted testing methodologies, which primarily rely on tester intuition, creativity, and adaptability. There are several tools and techniques that can aid testers in the exploratory testing process which we will be discussing in this talk.Tests of this kind are able to find bugs of this kind that are harder to find during structured testing or that other testing methods may have overlooked.The purpose of this abstract is to examine the nature and importance of exploratory testing in modern software development methods. It explores the fundamental ideas of exploratory testing, highlighting the value of domain knowledge and tester experience in spotting possible problems that may escape the notice of traditional testing methodologies. Throughout the software development lifecycle, exploratory testing promotes quick feedback loops and continuous improvement by giving testers the ability to make decisions in real time based on their observations. This abstract also clarifies the unique features of exploratory testing, like its non-linearity and capacity to replicate user behavior in real-world settings. Testers can find intricate bugs, usability problems, and edge cases in software through impromptu exploration that might go undetected. Exploratory testing's flexible and iterative structure fits in well with agile and DevOps processes, allowing for a quicker time to market without sacrificing the quality of the final product.Keywords: exploratory, testing, automation, quality
Procedia PDF Downloads 5145157 Impact of Emergency Medicine Department Crowding on Mortality
Authors: Morteza Gharibi, Abdolghader Pakniat, Somayeh Bahrampouri
Abstract:
Introduction: Emergency department (E.R.) crowding is a serious widespread problem in hospitals that leads to irregularities, a slower rate of delivery of services to patients, and a long-term stay. In addition, the long-term stay in the E.D. reduces the possibility of providing services with appropriate quality to other patients who are undergoing medical emergencies, which leads to dissatisfaction among patients. This study aimed to determine the relationship between ED-crowding and the mortality rate of the patients referred to the E.D. In a retrospective cohort study, all patients who expired in first 24 hours of admission were enrolled in the study. Crowding index at the moment of admission was calculated using Edwin Score. The data including history and physical examination, time of arrival in the E.D., diagnosis (using ICD 10 code), time of death, cause of death, demographic information was recoded based on triage forms on admission and patients’ medical files. Data analysis was performed by using descriptive statistics and chi square test, ANOVA tests using SPSS ver. 19. The time of arrival in E.D. to death in crowded E.D. conditions, with an average of five hours and 25 minutes, was significantly higher than the average admission Time of arrival in E.D. to death in active and crowded E.D. conditions. More physicians and nurses can be employed during crowded times to reduce staff fatigue and improve their performance during these hours.Keywords: mortality, emergency, department, crowding
Procedia PDF Downloads 9445156 Analysis in Mexico on Workers Performing Highly Repetitive Movements with Sensory Thermography in the Surface of the Wrist and Elbows
Authors: Sandra K. Enriquez, Claudia Camargo, Jesús E. Olguín, Juan A. López, German Galindo
Abstract:
Currently companies have increased the number of disorders of cumulative trauma (CTDs), these are increasing significantly due to the Highly Repetitive Movements (HRM) performed in workstations, which causes economic losses to businesses, due to temporary and permanent disabilities of workers. This analysis focuses on the prevention of disorders caused by: repeatability, duration and effort; And focuses on reducing cumulative trauma disorders such as occupational diseases using sensory thermography as a noninvasive method, the above is to evaluate the injuries could have workers to perform repetitive motions. Objectives: The aim is to define rest periods or job rotation before they generate a CTD, this sensory thermography by analyzing changes in temperature patterns on wrists and elbows when the worker is performing HRM over a period of time 2 hours and 30 minutes. Information on non-work variables such as wrist and elbow injuries, weight, gender, age, among others, and work variables such as temperature workspace, repetitiveness and duration also met. Methodology: The analysis to 4 industrial designers, 2 men and 2 women to be specific was conducted in a business in normal health for a period of 12 days, using the following time ranges: the first day for every 90 minutes continuous work were asked to rest 5 minutes, the second day for every 90 minutes of continuous work were asked to rest 10 minutes, the same to work 60 and 30 minutes straight. Each worker was tested with 6 different ranges at least twice. This analysis was performed in a controlled room temperature between 20 and 25 ° C, and a time to stabilize the temperature of the wrists and elbows than 20 minutes at the beginning and end of the analysis. Results: The range time of 90 minutes working continuous and a rest of 5 minutes of activity is where the maximum temperature (Tmax) was registered in the wrists and elbows in the office, we found the Tmax was 35.79 ° C with a difference of 2.79 ° C between the initial and final temperature of the left elbow presented at the individual 4 during the 86 minutes, in of range in 90 minutes continuously working and rested for 5 minutes of your activity. Conclusions: It is possible with this alternative technology is sensory thermography predict ranges of rotation or rest for the prevention of CTD to perform HRM work activities, obtaining with this reduce occupational disease, quotas by health agencies and increasing the quality of life of workers, taking this technology a cost-benefit acceptable in the future.Keywords: sensory thermography, temperature, cumulative trauma disorder (CTD), highly repetitive movement (HRM)
Procedia PDF Downloads 42945155 Spatiotemporal Analysis of Visual Evoked Responses Using Dense EEG
Authors: Rima Hleiss, Elie Bitar, Mahmoud Hassan, Mohamad Khalil
Abstract:
A comprehensive study of object recognition in the human brain requires combining both spatial and temporal analysis of brain activity. Here, we are mainly interested in three issues: the time perception of visual objects, the ability of discrimination between two particular categories (objects vs. animals), and the possibility to identify a particular spatial representation of visual objects. Our experiment consisted of acquiring dense electroencephalographic (EEG) signals during a picture-naming task comprising a set of objects and animals’ images. These EEG responses were recorded from nine participants. In order to determine the time perception of the presented visual stimulus, we analyzed the Event Related Potentials (ERPs) derived from the recorded EEG signals. The analysis of these signals showed that the brain perceives animals and objects with different time instants. Concerning the discrimination of the two categories, the support vector machine (SVM) was applied on the instantaneous EEG (excellent temporal resolution: on the order of millisecond) to categorize the visual stimuli into two different classes. The spatial differences between the evoked responses of the two categories were also investigated. The results showed a variation of the neural activity with the properties of the visual input. Results showed also the existence of a spatial pattern of electrodes over particular regions of the scalp in correspondence to their responses to the visual inputs.Keywords: brain activity, categorization, dense EEG, evoked responses, spatio-temporal analysis, SVM, time perception
Procedia PDF Downloads 42245154 Comparative Study of Conventional and Satellite Based Agriculture Information System
Authors: Rafia Hassan, Ali Rizwan, Sadaf Farhan, Bushra Sabir
Abstract:
The purpose of this study is to compare the conventional crop monitoring system with the satellite based crop monitoring system in Pakistan. This study is conducted for SUPARCO (Space and Upper Atmosphere Research Commission). The study focused on the wheat crop, as it is the main cash crop of Pakistan and province of Punjab. This study will answer the following: Which system is better in terms of cost, time and man power? The man power calculated for Punjab CRS is: 1,418 personnel and for SUPARCO: 26 personnel. The total cost calculated for SUPARCO is almost 13.35 million and CRS is 47.705 million. The man hours calculated for CRS (Crop Reporting Service) are 1,543,200 hrs (136 days) and man hours for SUPARCO are 8, 320hrs (40 days). It means that SUPARCO workers finish their work 96 days earlier than CRS workers. The results show that the satellite based crop monitoring system is efficient in terms of manpower, cost and time as compared to the conventional system, and also generates early crop forecasts and estimations. The research instruments used included: Interviews, physical visits, group discussions, questionnaires, study of reports and work flows. A total of 93 employees were selected using Yamane’s formula for data collection, which is done with the help questionnaires and interviews. Comparative graphing is used for the analysis of data to formulate the results of the research. The research findings also demonstrate that although conventional methods have a strong impact still in Pakistan (for crop monitoring) but it is the time to bring a change through technology, so that our agriculture will also be developed along modern lines.Keywords: area frame, crop reporting service, CRS, sample frame, SRS/GIS, satellite remote sensing/ geographic information system
Procedia PDF Downloads 29145153 A Framework for Information Quality in Accounting Information Systems Adoption
Authors: Wongsim Manirath
Abstract:
In order to implement AIS adoption successfully, it is important to consider the quality of information management and understand Information Quality (IQ) factors influencing AIS adoption. This research aims to explore ways of managing AIS adoption to investigate the adoption of accounting information systems within organisations. The study has led to the development of a framework for understanding the AIS adoption process in an organisation. This research used qualitative, interpretive evidence. This framework was developed from case studies and by collecting qualitative data (interviews). This research has conducted 10 case studies to study how IQ is managed through the accounting information system adoption process. A special focus is placed on determining how organisation size influences the information quality practices. The finding is especially useful to SMEs as many SMEs have the desire to grow bigger. By better dealing with IQ issues, there could be a successful future.Keywords: data quality, information quality, accounting information system, information management
Procedia PDF Downloads 46845152 Navigating Construction Project Outcomes: Synergy Through the Evolution of Digital Innovation and Strategic Management
Authors: Derrick Mirindi, Frederic Mirindi, Oluwakemi Oshineye
Abstract:
The ongoing high rate of construction project failures worldwide is often blamed on the difficulties of managing stakeholders. This highlights the crucial role of strategic management (SM) in achieving project success. This study investigates how integrating digital tools into the SM framework can effectively address stakeholder-related challenges. This work specifically focuses on the impact of evolving digital tools, such as Project Management Software (PMS) (e.g., Basecamp and Wrike), Building Information Modeling (BIM) (e.g., Tekla BIMsight and Autodesk Navisworks), Virtual and Augmented Reality (VR/AR) (e.g., Microsoft HoloLens), drones and remote monitoring, and social media and Web-Based platforms, in improving stakeholder engagement and project outcomes. Through existing literature with examples of failed projects, the study highlights how the evolution of digital tools will serve as facilitators within the strategic management process. These tools offer benefits such as real-time data access, enhanced visualization, and more efficient workflows to mitigate stakeholder challenges in construction projects. The findings indicate that integrating digital tools with SM principles effectively addresses stakeholder challenges, resulting in improved project outcomes and stakeholder satisfaction. The research advocates for a combined approach that embraces both strategic management and digital innovation to navigate the complex stakeholder landscape in construction projects.Keywords: strategic management, digital tools, virtual and augmented reality, stakeholder management, building information modeling, project management software
Procedia PDF Downloads 8345151 Random Access in IoT Using Naïve Bayes Classification
Authors: Alhusein Almahjoub, Dongyu Qiu
Abstract:
This paper deals with the random access procedure in next-generation networks and presents the solution to reduce total service time (TST) which is one of the most important performance metrics in current and future internet of things (IoT) based networks. The proposed solution focuses on the calculation of optimal transmission probability which maximizes the success probability and reduces TST. It uses the information of several idle preambles in every time slot, and based on it, it estimates the number of backlogged IoT devices using Naïve Bayes estimation which is a type of supervised learning in the machine learning domain. The estimation of backlogged devices is necessary since optimal transmission probability depends on it and the eNodeB does not have information about it. The simulations are carried out in MATLAB which verify that the proposed solution gives excellent performance.Keywords: random access, LTE/LTE-A, 5G, machine learning, Naïve Bayes estimation
Procedia PDF Downloads 14545150 Improving Law Enforcement Strategies Through Geographic Information Systems: A Spatio-Temporal Analysis of Antisocial Activities in Móstoles (2022)
Authors: Daniel Suarez Alonso
Abstract:
This study has tried to focus on the alternatives offered to police institutions by the implementation of Geographic Information systems. Providing operational police commanders with effective and efficient tools, providing analytical capacity to reduce criminal opportunities, must be a priority. Given the intimate connection of crimes and infractions to the environment, law enforcement institutions must respond proactively to changing circumstances of anti-norm behaviors. To this end, it has been intended to analyze the antisocial spatial distribution of the city of Móstoles, trying to identify those spatiotemporal patterns that occur to anticipate their commission through the planning of dynamic preventive strategies. The application of GIS offers alternative analytical approaches to the different problems that underlie the development of life in society, focusing resources on those places with the highest concentration of incidents.Keywords: data analysis, police organizations, police prevention, geographic information systems
Procedia PDF Downloads 5045149 Gan Nanowire-Based Sensor Array for the Detection of Cross-Sensitive Gases Using Principal Component Analysis
Authors: Ashfaque Hossain Khan, Brian Thomson, Ratan Debnath, Abhishek Motayed, Mulpuri V. Rao
Abstract:
Though the efforts had been made, the problem of cross-sensitivity for a single metal oxide-based sensor can’t be fully eliminated. In this work, a sensor array has been designed and fabricated comprising of platinum (Pt), copper (Cu), and silver (Ag) decorated TiO2 and ZnO functionalized GaN nanowires using industry-standard top-down fabrication approach. The metal/metal-oxide combinations within the array have been determined from prior molecular simulation study using first principle calculations based on density functional theory (DFT). The gas responses were obtained for both single and mixture of NO2, SO2, ethanol, and H2 in the presence of H2O and O2 gases under UV light at room temperature. Each gas leaves a unique response footprint across the array sensors by which precise discrimination of cross-sensitive gases has been achieved. An unsupervised principal component analysis (PCA) technique has been implemented on the array response. Results indicate that each gas forms a distinct cluster in the score plot for all the target gases and their mixtures, indicating a clear separation among them. In addition, the developed array device consumes very low power because of ultra-violet (UV) assisted sensing as compared to commercially available metal-oxide sensors. The nanowire sensor array, in combination with PCA, is a potential approach for precise real-time gas monitoring applications.Keywords: cross-sensitivity, gas sensor, principle component analysis (PCA), sensor array
Procedia PDF Downloads 10745148 Effect of Particle Shape on Monotonic and Cyclic Biaxial Behaviour of Sand Using Discrete Element Method
Authors: Raj Banerjee, Y. M. Parulekar, Aniruddha Sengupta, J. Chattopadhyay
Abstract:
This study proposes a Discrete Element Method (DEM) simulation using a commercial software PFC 2D (2019) for quantitatively simulating the monotonic and cyclic behaviour of sand using irregular shapes of sand grains. A preliminary analysis of the number of particles for optimal Representative Element Volume (REV) simulation of dimension 35mm x 35mm x 70mm using the scaled Grain Size Distribution (GSD) of sand is carried out. Subsequently, the effect of particle shape on the performance of sand during monotonic and cyclic bi-axial tests is assessed using numerical simulation. The validation of the numerical simulation for one case is carried out using the test results from the literature. Further numerical studies are performed in which the particles in REV are simulated by mixing round discs with irregular clumps (100% round disc, 75% round disc 25% irregular clump, 50% round disc 50% irregular clump, 25% round disc 75% irregular clump, 100% irregular clump) in different proportions using Dry Deposition (DD) method. The macro response for monotonic loading shows that irregular sand has a higher strength than round particles and that the Mohr-Coulomb failure envelope depends on the shape of the grains. During cyclic loading, it is observed that the liquefaction resistance curve (Cyclic Stress Ratio (CSR)-Number of cycles (N)) of sand is dependent on the combination of particle shapes with different proportions.Keywords: biaxial test, particle shape, monotonic, cyclic
Procedia PDF Downloads 7245147 The Effect of Manure Loaded Biochar on Soil Microbial Communities
Authors: T. Weber, D. MacKenzie
Abstract:
The script in this paper describes the use of advanced simulation environment using electronic systems (microcontroller, operational amplifiers, and FPGA). The simulation was used for non-linear dynamic systems behaviour with required observer structure working with parallel real-time simulation based on state-space representation. The proposed deposited model was used for electrodynamic effects including ionising effects and eddy current distribution also. With the script and proposed method, it is possible to calculate the spatial distribution of the electromagnetic fields in real-time and such systems. For further purpose, the spatial temperature distribution may also be used. With upon system, the uncertainties and disturbances may be determined. This provides the estimation of the more precise system states for the required system and additionally the estimation of the ionising disturbances that arise due to radiation effects in space systems. The results have also shown that a system can be developed specifically with the real-time calculation (estimation) of the radiation effects only. Electronic systems can take damage caused by impacts with charged particle flux in space or radiation environment. TID (Total Ionising Dose) of 1 Gy and Single Effect Transient (SET) free operation up to 50 MeVcm²/mg may assure certain functions. Single-Event Latch-up (SEL) results on the placement of several transistors in the shared substrate of an integrated circuit; ionising radiation can activate an additional parasitic thyristor. This short circuit between semiconductor-elements can destroy the device without protection and measurements. Single-Event Burnout (SEB) on the other hand, increases current between drain and source of a MOSFET and destroys the component in a short time. A Single-Event Gate Rupture (SEGR) can destroy a dielectric of semiconductor also. In order to be able to react to these processes, it must be calculated within a shorter time that ionizing radiation and dose is present. For this purpose, sensors may be used for the realistic evaluation of the diffusion and ionizing effects of the test system. For this purpose, the Peltier element is used for the evaluation of the dynamic temperature increases (dT/dt), from which a measure of the ionization processes and thus radiation will be detected. In addition, the piezo element may be used to record highly dynamic vibrations and oscillations to absorb impacts of charged particle flux. All available sensors shall be used to calibrate the spatial distributions also. By measured value of size and known location of the sensors, the entire distribution in space can be calculated retroactively or more accurately. With the formation, the type of ionisation and the direct effect to the systems and thus possible prevent processes can be activated up to the shutdown. The results show possibilities to perform more qualitative and faster simulations independent of space-systems and radiation environment also. The paper gives additionally an overview of the diffusion effects and their mechanisms.Keywords: cattle, biochar, manure, microbial activity
Procedia PDF Downloads 10345146 Message Authentication Scheme for Vehicular Ad-Hoc Networks under Sparse RSUs Environment
Authors: Wen Shyong Hsieh, Chih Hsueh Lin
Abstract:
In this paper, we combine the concepts of chameleon hash function (CHF) and identification based cryptography (IBC) to build a message authentication environment for VANET under sparse RSUs. Based on the CHF, TA keeps two common secrets that will be embedded to all identities to be as the evidence of mutual trusting. TA will issue one original identity to every RSU and vehicle. An identity contains one public ID and one private key. The public ID, includes three components: pseudonym, random key, and public key, is used to present one entity and can be verified to be a legal one. The private key is used to claim the ownership of the public ID. Based on the concept of IBC, without any negotiating process, a CHF pairing key multiplied by one private key and other’s public key will be used for mutually trusting and to be utilized as the session key of secure communicating between RSUs and vehicles. To help the vehicles to do message authenticating, the RSUs are assigned to response the vehicle’s temple identity request using two short time secretes that are broadcasted by TA. To light the loading of request information, one day is divided into M time slots. At every time slot, TA will broadcast two short time secretes to all valid RSUs for that time slot. Any RSU can response the temple identity request from legal vehicles. With the collected announcement of public IDs from the neighbor vehicles, a vehicle can set up its neighboring set, which includes the information about the neighbor vehicle’s temple public ID and temple CHF pairing key that can be derived by the private key and neighbor’s public key and will be used to do message authenticating or secure communicating without the help of RSU.Keywords: Internet of Vehicles (IOV), Vehicular Ad-hoc Networks (VANETs), Chameleon Hash Function (CHF), message authentication
Procedia PDF Downloads 39145145 The Visualization of the Way of Creating a Service: Slavic Liturgical Books. Between Text and Music
Authors: Victoria Legkikh
Abstract:
To create a new Orthodox service of Jerusalem rite and to make it possible for a performance, one had to use several types of books. These are menaions and triodion, cleargy service book, stichirarion and typikon. These books keep a part of the information about the service, which a medieval copyist had to put together like a puzzle. But an abundance of necessary books and their variety created a lot of problems in copying services. The main problem was the difference of text in notated and not notated manuscripts (they were corrected at a different time) and lack of information in typikon, which provided only a type of hymns and their mode. After all, a copyist could have both corrected and not corrected manuscripts which also provided a different type of service. It brings us to the situation when we hardly have a couple of manuscripts containing the same service, and it is difficult to understand which changes were made voluntarily and which ones were provided by different types of available manuscripts. A recent paper proposes an analysis of every type of liturgical book and a way of using them in copying and correcting a service so we can divide voluntary changes and changes due to various types of books. The paper also proposes an index showing the “material” life of hymns in different types of manuscripts and the changes of its version and place in the same type of manuscript. This type of index can help in reconstructing the way of creation/copying service and can be useful for publication of the services providing necessary information of every hymn in every used manuscript.Keywords: orthodox church music, creation, manuscripts, liturgical books
Procedia PDF Downloads 17345144 Comparison of Parametric and Bayesian Survival Regression Models in Simulated and HIV Patient Antiretroviral Therapy Data: Case Study of Alamata Hospital, North Ethiopia
Authors: Zeytu G. Asfaw, Serkalem K. Abrha, Demisew G. Degefu
Abstract:
Background: HIV/AIDS remains a major public health problem in Ethiopia and heavily affecting people of productive and reproductive age. We aimed to compare the performance of Parametric Survival Analysis and Bayesian Survival Analysis using simulations and in a real dataset application focused on determining predictors of HIV patient survival. Methods: A Parametric Survival Models - Exponential, Weibull, Log-normal, Log-logistic, Gompertz and Generalized gamma distributions were considered. Simulation study was carried out with two different algorithms that were informative and noninformative priors. A retrospective cohort study was implemented for HIV infected patients under Highly Active Antiretroviral Therapy in Alamata General Hospital, North Ethiopia. Results: A total of 320 HIV patients were included in the study where 52.19% females and 47.81% males. According to Kaplan-Meier survival estimates for the two sex groups, females has shown better survival time in comparison with their male counterparts. The median survival time of HIV patients was 79 months. During the follow-up period 89 (27.81%) deaths and 231 (72.19%) censored individuals registered. The average baseline cluster of differentiation 4 (CD4) cells count for HIV/AIDS patients were 126.01 but after a three-year antiretroviral therapy follow-up the average cluster of differentiation 4 (CD4) cells counts were 305.74, which was quite encouraging. Age, functional status, tuberculosis screen, past opportunistic infection, baseline cluster of differentiation 4 (CD4) cells, World Health Organization clinical stage, sex, marital status, employment status, occupation type, baseline weight were found statistically significant factors for longer survival of HIV patients. The standard error of all covariate in Bayesian log-normal survival model is less than the classical one. Hence, Bayesian survival analysis showed better performance than classical parametric survival analysis, when subjective data analysis was performed by considering expert opinions and historical knowledge about the parameters. Conclusions: Thus, HIV/AIDS patient mortality rate could be reduced through timely antiretroviral therapy with special care on the potential factors. Moreover, Bayesian log-normal survival model was preferable than the classical log-normal survival model for determining predictors of HIV patients survival.Keywords: antiretroviral therapy (ART), Bayesian analysis, HIV, log-normal, parametric survival models
Procedia PDF Downloads 19645143 Structural Properties of Polar Liquids in Binary Mixture Using Microwave Technique
Authors: Shagufta Tabassum, V. P. Pawar
Abstract:
The study of static dielectric properties in a binary mixture of 1,2 dichloroethane (DE) and n,n dimethylformamide (DMF) polar liquids has been carried out in the frequency range of 10 MHz to 30 GHz for 11 different concentration using time domain reflectometry technique at 10ºC temperature. The dielectric relaxation study of solute-solvent mixture at microwave frequencies gives information regarding the creation of monomers and multimers as well as interaction between the molecules of the binary mixture. The least squares fit method is used to determine the values of dielectric parameters such as static dielectric constant (ε0), dielectric constant at high frequency (ε∞) and relaxation time (τ).Keywords: shagufta shaikhexcess parameters, relaxation time, static dielectric constant, time domain reflectometry
Procedia PDF Downloads 24445142 Emerging Threats and Adaptive Defenses: Navigating the Future of Cybersecurity in a Hyperconnected World
Authors: Olasunkanmi Jame Ayodeji, Adebayo Adeyinka Victor
Abstract:
In a hyperconnected world, cybersecurity faces a continuous evolution of threats that challenge traditional defence mechanisms. This paper explores emerging cybersecurity threats like malware, ransomware, phishing, social engineering, and the Internet of Things (IoT) vulnerabilities. It delves into the inadequacies of existing cybersecurity defences in addressing these evolving risks and advocates for adaptive defence mechanisms that leverage AI, machine learning, and zero-trust architectures. The paper proposes collaborative approaches, including public-private partnerships and information sharing, as essential to building a robust defence strategy to address future cyber threats. The need for continuous monitoring, real-time incident response, and adaptive resilience strategies is highlighted to fortify digital infrastructures in the face of escalating global cyber risks.Keywords: cybersecurity, hyperconnectivity, malware, adaptive defences, zero-trust architecture, internet of things vulnerabilities
Procedia PDF Downloads 2245141 Analysis of DNA from Fired Cartridge Casings
Authors: S. Mawlood, L. Denanny, N. Watson, B. Pickard
Abstract:
DNA analysis has been widely accepted as providing valuable evidence concerning the identity of the source of biological traces. Our work has showed that DNA samples can survive on cartridges even after firing. The study also raised the possibility of determining other information such as the age of the donor. Such information may be invaluable in certain cases where spent cartridges from automatic weapons are left behind at the scene of a crime. In spite of the nature of touch evidence and exposure to high chamber temperatures during shooting, we were still capable to retrieve enough DNA for profile typing. In order to estimate age of contributor, DNA methylation levels were analyzed using EpiTect system for retrieved DNA. However, results were not conclusive, due to low amount of input DNA.Keywords: DNA profile, DNA Methylation, fired cartridge, touch sample
Procedia PDF Downloads 45245140 Women’s Colours in Digital Innovation
Authors: Daniel J. Patricio Jiménez
Abstract:
Digital reality demands new ways of thinking, flexibility in learning, acquisition of new competencies, visualizing reality under new approaches, generating open spaces, understanding dimensions in continuous change, etc. We need inclusive growth, where colors are not lacking, where lights do not give a distorted reality, where science is not half-truth. In carrying out this study, the documentary or bibliographic collection has been taken into account, providing a reflective and analytical analysis of current reality. In this context, deductive and inductive methods have been used on different multidisciplinary information sources. Women today and tomorrow are a strategic element in science and arts, which, under the umbrella of sustainability, implies ‘meeting current needs without detriment to future generations’. We must build new scenarios, which qualify ‘the feminine and the masculine’ as an inseparable whole, encouraging cooperative behavior; nothing is exclusive or excluding, and that is where true respect for diversity must be based. We are all part of an ecosystem, which we will make better as long as there is a real balance in terms of gender. It is the time of ‘the lifting of the veil’, in other words, it is the time to discover the pseudonyms, the women who painted, wrote, investigated, recorded advances, etc. However, the current reality demands much more; we must remove doors where they are not needed. Mass processing of data, big data, needs to incorporate algorithms under the perspective of ‘the feminine’. However, most STEM students (science, technology, engineering, and math) are men. Our way of doing science is biased, focused on honors and short-term results to the detriment of sustainability. Historically, the canons of beauty, the way of looking, of perceiving, of feeling, depended on the circumstances and interests of each moment, and women had no voice in this. Parallel to science, there is an under-representation of women in the arts, but not so much in the universities, but when we look at galleries, museums, art dealers, etc., colours impoverish the gaze and once again highlight the gender gap and the silence of the feminine. Art registers sensations by divining the future, science will turn them into reality. The uniqueness of the so-called new normality requires women to be protagonists both in new forms of emotion and thought, and in the experimentation and development of new models. This will result in women playing a decisive role in the so-called "5.0 society" or, in other words, in a more sustainable, more humane world.Keywords: art, digitalization, gender, science
Procedia PDF Downloads 16545139 A Hybrid Algorithm for Collaborative Transportation Planning among Carriers
Authors: Elham Jelodari Mamaghani, Christian Prins, Haoxun Chen
Abstract:
In this paper, there is concentration on collaborative transportation planning (CTP) among multiple carriers with pickup and delivery requests and time windows. This problem is a vehicle routing problem with constraints from standard vehicle routing problems and new constraints from a real-world application. In the problem, each carrier has a finite number of vehicles, and each request is a pickup and delivery request with time window. Moreover, each carrier has reserved requests, which must be served by itself, whereas its exchangeable requests can be outsourced to and served by other carriers. This collaboration among carriers can help them to reduce total transportation costs. A mixed integer programming model is proposed to the problem. To solve the model, a hybrid algorithm that combines Genetic Algorithm and Simulated Annealing (GASA) is proposed. This algorithm takes advantages of GASA at the same time. After tuning the parameters of the algorithm with the Taguchi method, the experiments are conducted and experimental results are provided for the hybrid algorithm. The results are compared with those obtained by a commercial solver. The comparison indicates that the GASA significantly outperforms the commercial solver.Keywords: centralized collaborative transportation, collaborative transportation with pickup and delivery, collaborative transportation with time windows, hybrid algorithm of GA and SA
Procedia PDF Downloads 39245138 Joint Training Offer Selection and Course Timetabling Problems: Models and Algorithms
Authors: Gianpaolo Ghiani, Emanuela Guerriero, Emanuele Manni, Alessandro Romano
Abstract:
In this article, we deal with a variant of the classical course timetabling problem that has a practical application in many areas of education. In particular, in this paper we are interested in high schools remedial courses. The purpose of such courses is to provide under-prepared students with the skills necessary to succeed in their studies. In particular, a student might be under prepared in an entire course, or only in a part of it. The limited availability of funds, as well as the limited amount of time and teachers at disposal, often requires schools to choose which courses and/or which teaching units to activate. Thus, schools need to model the training offer and the related timetabling, with the goal of ensuring the highest possible teaching quality, by meeting the above-mentioned financial, time and resources constraints. Moreover, there are some prerequisites between the teaching units that must be satisfied. We first present a Mixed-Integer Programming (MIP) model to solve this problem to optimality. However, the presence of many peculiar constraints contributes inevitably in increasing the complexity of the mathematical model. Thus, solving it through a general purpose solver may be performed for small instances only, while solving real-life-sized instances of such model requires specific techniques or heuristic approaches. For this purpose, we also propose a heuristic approach, in which we make use of a fast constructive procedure to obtain a feasible solution. To assess our exact and heuristic approaches we perform extensive computational results on both real-life instances (obtained from a high school in Lecce, Italy) and randomly generated instances. Our tests show that the MIP model is never solved to optimality, with an average optimality gap of 57%. On the other hand, the heuristic algorithm is much faster (in about the 50% of the considered instances it converges in approximately half of the time limit) and in many cases allows achieving an improvement on the objective function value obtained by the MIP model. Such an improvement ranges between 18% and 66%.Keywords: heuristic, MIP model, remedial course, school, timetabling
Procedia PDF Downloads 60545137 Pattern of Stress Distribution in Different Ligature-Wire-Brackets Systems: A FE and Experimental Analysis
Authors: Afef Dridi, Salah Mezlini
Abstract:
Since experimental devices cannot calculate stress and deformation of complex structures. The Finite Element Method FEM has been widely used in several fields of research. One of these fields is orthodontics. The advantage of using such a method is the use of an accurate and non invasive method that allows us to have a sufficient data about the physiological reactions can happening in soft tissues. Most of researches done in this field were interested in the study of stresses and deformations induced by orthodontic apparatus in soft tissues (alveolar tissues). Only few studies were interested in the distribution of stress and strain in the orthodontic brackets. These studies, although they tried to be as close as possible to real conditions, their models did not reproduce the clinical cases. For this reason, the model generated by our research is the closest one to reality. In this study, a numerical model was developed to explore the stress and strain distribution under the application of real conditions. A comparison between different material properties was also done.Keywords: visco-hyperelasticity, FEM, orthodontic treatment, inverse method
Procedia PDF Downloads 25945136 Use of Socially Assistive Robots in Early Rehabilitation to Promote Mobility for Infants with Motor Delays
Authors: Elena Kokkoni, Prasanna Kannappan, Ashkan Zehfroosh, Effrosyni Mavroudi, Kristina Strother-Garcia, James C. Galloway, Jeffrey Heinz, Rene Vidal, Herbert G. Tanner
Abstract:
Early immobility affects the motor, cognitive, and social development. Current pediatric rehabilitation lacks the technology that will provide the dosage needed to promote mobility for young children at risk. The addition of socially assistive robots in early interventions may help increase the mobility dosage. The aim of this study is to examine the feasibility of an early intervention paradigm where non-walking infants experience independent mobility while socially interacting with robots. A dynamic environment is developed where both the child and the robot interact and learn from each other. The environment involves: 1) a range of physical activities that are goal-oriented, age-appropriate, and ability-matched for the child to perform, 2) the automatic functions that perceive the child’s actions through novel activity recognition algorithms, and decide appropriate actions for the robot, and 3) a networked visual data acquisition system that enables real-time assessment and provides the means to connect child behavior with robot decision-making in real-time. The environment was tested by bringing a two-year old boy with Down syndrome for eight sessions. The child presented delays throughout his motor development with the current being on the acquisition of walking. During the sessions, the child performed physical activities that required complex motor actions (e.g. climbing an inclined platform and/or staircase). During these activities, a (wheeled or humanoid) robot was either performing the action or was at its end point 'signaling' for interaction. From these sessions, information was gathered to develop algorithms to automate the perception of activities which the robot bases its actions on. A Markov Decision Process (MDP) is used to model the intentions of the child. A 'smoothing' technique is used to help identify the model’s parameters which are a critical step when dealing with small data sets such in this paradigm. The child engaged in all activities and socially interacted with the robot across sessions. With time, the child’s mobility was increased, and the frequency and duration of complex and independent motor actions were also increased (e.g. taking independent steps). Simulation results on the combination of the MDP and smoothing support the use of this model in human-robot interaction. Smoothing facilitates learning MDP parameters from small data sets. This paradigm is feasible and provides an insight on how social interaction may elicit mobility actions suggesting a new early intervention paradigm for very young children with motor disabilities. Acknowledgment: This work has been supported by NIH under grant #5R01HD87133.Keywords: activity recognition, human-robot interaction, machine learning, pediatric rehabilitation
Procedia PDF Downloads 29245135 Evaluating Construction Project Outcomes: Synergy Through the Evolution of Digital Innovation and Strategic Management
Authors: Mirindi Derrick, Mirindi Frederic, Oluwakemi Oshineye
Abstract:
Abstract: The ongoing high rate of construction project failures worldwide is often blamed on the difficulties of managing stakeholders. This highlights the crucial role of strategic management (SM) in achieving project success. This study investigates how integrating digital tools into the SM framework can effectively address stakeholder-related challenges. This work specifically focuses on the impact of evolving digital tools, such as Project Management Software (PMS) (e.g., Basecamp and Wrike), Building Information Modeling (BIM) (e.g., Tekla BIMsight and Autodesk Navisworks), Virtual and Augmented Reality (VR/AR) (e.g., Microsoft HoloLens), drones and remote monitoring, and social media and Web-Based platforms, in improving stakeholder engagement and project outcomes. Through existing literature with examples of failed projects, the study highlights how the evolution of digital tools will serve as facilitators within the strategic management process. These tools offer benefits such as real-time data access, enhanced visualization, and more efficient workflows to mitigate stakeholder challenges in construction projects. The findings indicate that integrating digital tools with SM principles effectively addresses stakeholder challenges, resulting in improved project outcomes and stakeholder satisfaction. The research advocates for a combined approach that embraces both strategic management and digital innovation to navigate the complex stakeholder landscape in construction projects.Keywords: strategic management, digital tools, virtual and augmented reality, stakeholder management, building information modeling, project management software
Procedia PDF Downloads 50