Search results for: data mining applications and discovery
27184 Tackling the Value-Action-Gap: Improving Civic Participation Using a Holistic Behavioral Model Approach
Authors: Long Pham, Julia Blanke
Abstract:
An increasingly popular way of establishing citizen engagement within communities is through ‘city apps’. Currently, most of these mobile applications seem to be extensions of the existing communication media, sometimes merely replicating the information available on the classical city web sites, and therefore provide minimal additional impact on citizen behavior and engagement. In order to overcome this challenge, we propose to use a holistic behavioral model to generate dynamic and contextualized app content based on optimizing well defined city-related performance goals constrained by the proposed behavioral model. In this paper, we will show how the data collected by the CorkCitiEngage project in the Irish city of Cork can be utilized to calibrate aspects of the proposed model enabling the design of a personalized citizen engagement app aiming at positively influencing people’s behavior towards more active participation in their communities. We will focus on the important aspect of intentions to act, which is essential for understanding the reasons behind the common value-action-gap being responsible for the mismatch between good intentions and actual observable behavior, and will discuss how customized app design can be based on a rigorous model of behavior optimized towards maximizing well defined city-related performance goals.Keywords: city apps, holistic behaviour model, intention to act, value-action-gap, citizen engagement
Procedia PDF Downloads 23027183 Simulations to Predict Solar Energy Potential by ERA5 Application at North Africa
Authors: U. Ali Rahoma, Nabil Esawy, Fawzia Ibrahim Moursy, A. H. Hassan, Samy A. Khalil, Ashraf S. Khamees
Abstract:
The design of any solar energy conversion system requires the knowledge of solar radiation data obtained over a long period. Satellite data has been widely used to estimate solar energy where no ground observation of solar radiation is available, yet there are limitations on the temporal coverage of satellite data. Reanalysis is a “retrospective analysis” of the atmosphere parameters generated by assimilating observation data from various sources, including ground observation, satellites, ships, and aircraft observation with the output of NWP (Numerical Weather Prediction) models, to develop an exhaustive record of weather and climate parameters. The evaluation of the performance of reanalysis datasets (ERA-5) for North Africa against high-quality surface measured data was performed using statistical analysis. The estimation of global solar radiation (GSR) distribution over six different selected locations in North Africa during ten years from the period time 2011 to 2020. The root means square error (RMSE), mean bias error (MBE) and mean absolute error (MAE) of reanalysis data of solar radiation range from 0.079 to 0.222, 0.0145 to 0.198, and 0.055 to 0.178, respectively. The seasonal statistical analysis was performed to study seasonal variation of performance of datasets, which reveals the significant variation of errors in different seasons—the performance of the dataset changes by changing the temporal resolution of the data used for comparison. The monthly mean values of data show better performance, but the accuracy of data is compromised. The solar radiation data of ERA-5 is used for preliminary solar resource assessment and power estimation. The correlation coefficient (R2) varies from 0.93 to 99% for the different selected sites in North Africa in the present research. The goal of this research is to give a good representation for global solar radiation to help in solar energy application in all fields, and this can be done by using gridded data from European Centre for Medium-Range Weather Forecasts ECMWF and producing a new model to give a good result.Keywords: solar energy, solar radiation, ERA-5, potential energy
Procedia PDF Downloads 21627182 Vertical Accuracy Evaluation of Indian National DEM (CartoDEM v3) Using Dual Frequency GNSS Derived Ground Control Points for Lower Tapi Basin, Western India
Authors: Jaypalsinh B. Parmar, Pintu Nakrani, Ashish Chaurasia
Abstract:
Digital Elevation Model (DEM) is considered as an important data in GIS-based terrain analysis for many applications and assessment of processes such as environmental and climate change studies, hydrologic modelling, etc. Vertical accuracy of DEM having geographically dynamic nature depends on different parameters which affect the model simulation outcomes. Vertical accuracy assessment in Indian landscape especially in low-lying coastal urban terrain such as lower Tapi Basin is very limited. In the present study, attempt has been made to evaluate the vertical accuracy of 30m resolution open source Indian National Cartosat-1 DEM v3 for Lower Tapi Basin (LTB) from western India. The extensive field investigation is carried out using stratified random fast static DGPS survey in the entire study region, and 117 high accuracy ground control points (GCPs) have been obtained. The above open source DEM was compared with obtained GCPs, and different statistical attributes were envisaged, and vertical error histograms were also evaluated.Keywords: CartoDEM, Digital Elevation Model, GPS, lower Tapi basin
Procedia PDF Downloads 36227181 Indian Road Traffic Flow Analysis Using Blob Tracking from Video Sequences
Authors: Balaji Ganesh Rajagopal, Subramanian Appavu alias Balamurugan, Ayyalraj Midhun Kumar, Krishnan Nallaperumal
Abstract:
Intelligent Transportation System is an Emerging area to solve multiple transportation problems. Several forms of inputs are needed in order to solve ITS problems. Advanced Traveler Information System (ATIS) is a core and important ITS area of this modern era. This involves travel time forecasting, efficient road map analysis and cost based path selection, Detection of the vehicle in the dynamic conditions and Traffic congestion state forecasting. This Article designs and provides an algorithm for traffic data generation which can be used for the above said ATIS application. By inputting the real world traffic situation in the form of video sequences, the algorithm determines the Traffic density in terms of congestion, number of vehicles in a given path which can be fed for various ATIS applications. The Algorithm deduces the key frame from the video sequences and follows the Blob detection, Identification and Tracking using connected components algorithm to determine the correlation between the vehicles moving in the real road scene.Keywords: traffic transportation, traffic density estimation, blob identification and tracking, relative velocity of vehicles, correlation between vehicles
Procedia PDF Downloads 51427180 Efficient Pre-Processing of Single-Cell Assay for Transposase Accessible Chromatin with High-Throughput Sequencing Data
Authors: Fan Gao, Lior Pachter
Abstract:
The primary tool currently used to pre-process 10X Chromium single-cell ATAC-seq data is Cell Ranger, which can take very long to run on standard datasets. To facilitate rapid pre-processing that enables reproducible workflows, we present a suite of tools called scATAK for pre-processing single-cell ATAC-seq data that is 15 to 18 times faster than Cell Ranger on mouse and human samples. Our tool can also calculate chromatin interaction potential matrices, and generate open chromatin signal and interaction traces for cell groups. We use scATAK tool to explore the chromatin regulatory landscape of a healthy adult human brain and unveil cell-type specific features, and show that it provides a convenient and computational efficient approach for pre-processing single-cell ATAC-seq data.Keywords: single-cell, ATAC-seq, bioinformatics, open chromatin landscape, chromatin interactome
Procedia PDF Downloads 15827179 Enabling Cloud Adoption Based Secured Mobile Banking through Backend as a Service
Authors: P. S. Jagadeesh Kumar, S. Meenakshi Sundaram
Abstract:
With the increase of prevailing non-traditional rivalry, mobile banking experiences an ever changing commercial backdrop. Substantial customer demands have established to be more intricate as customers request more expediency and superintend over their banking services. To enterprise advance and modernization in mobile banking applications, it is gradually obligatory to deeply leapfrog the scuffle using business model transformation. The dramaturgical vicissitudes taking place in mobile banking entail advanced traditions to exploit security. By reforming and transforming older back office into integrated mobile banking applications, banks can engender a supple and nimble banking environment that can rapidly respond to new business requirements over cloud computing. Cloud computing is transfiguring ecosystems in numerous industries, and mobile banking is no exemption providing services innovation, greater flexibility to respond to improved security and enhanced business intelligence with less cost. Cloud technology offer secure deployment possibilities that can provision banks in developing new customer experiences, empower operative relationship and advance speed to efficient banking transaction. Cloud adoption is escalating quickly since it can be made secured for commercial mobile banking transaction through backend as a service in scrutinizing the security strategies of the cloud service provider along with the antiquity of transaction details and their security related practices.Keywords: cloud adoption, backend as a service, business intelligence, secured mobile banking
Procedia PDF Downloads 25727178 Meta Mask Correction for Nuclei Segmentation in Histopathological Image
Authors: Jiangbo Shi, Zeyu Gao, Chen Li
Abstract:
Nuclei segmentation is a fundamental task in digital pathology analysis and can be automated by deep learning-based methods. However, the development of such an automated method requires a large amount of data with precisely annotated masks which is hard to obtain. Training with weakly labeled data is a popular solution for reducing the workload of annotation. In this paper, we propose a novel meta-learning-based nuclei segmentation method which follows the label correction paradigm to leverage data with noisy masks. Specifically, we design a fully conventional meta-model that can correct noisy masks by using a small amount of clean meta-data. Then the corrected masks are used to supervise the training of the segmentation model. Meanwhile, a bi-level optimization method is adopted to alternately update the parameters of the main segmentation model and the meta-model. Extensive experimental results on two nuclear segmentation datasets show that our method achieves the state-of-the-art result. In particular, in some noise scenarios, it even exceeds the performance of training on supervised data.Keywords: deep learning, histopathological image, meta-learning, nuclei segmentation, weak annotations
Procedia PDF Downloads 14327177 Simulation and Experimental Research on Pocketing Operation for Toolpath Optimization in CNC Milling
Authors: Rakesh Prajapati, Purvik Patel, Avadhoot Rajurkar
Abstract:
Nowadays, manufacturing industries augment their production lines with modern machining centers backed by CAM software. Several attempts are being made to cut down the programming time for machining complex geometries. Special programs/software have been developed to generate the digital numerical data and to prepare NC programs by using suitable post-processors for different machines. By selecting the tools and manufacturing process then applying tool paths and NC program are generated. More and more complex mechanical parts that earlier were being cast and assembled/manufactured by other processes are now being machined. Majority of these parts require lots of pocketing operations and find their applications in die and mold, turbo machinery, aircraft, nuclear, defense etc. Pocketing operations involve removal of large quantity of material from the metal surface. The modeling of warm cast and clamping a piece of food processing parts which the used of Pro-E and MasterCAM® software. Pocketing operation has been specifically chosen for toolpath optimization. Then after apply Pocketing toolpath, Multi Tool Selection and Reduce Air Time give the results of software simulation time and experimental machining time.Keywords: toolpath, part program, optimization, pocket
Procedia PDF Downloads 28827176 Swelling Hydrogels on the Base Nitron Fiber Wastes for Water Keeping in Sandy Soils
Authors: Alim Asamatdinov
Abstract:
Superabsorbent polymer hydrogels can swell to absorb huge volumes of water or aqueous solutions. This property has led to many practical applications of these new materials, particularly in agriculture for improving the water retention of soils and the water supply of plants. This article reviews the methods of polymeric hydrogels, measurements and treatments of their properties, as well as their effects on soil and on plant growth. The thermodynamic approach used to describe the swelling behaviour of polymer networks proves to be quite helpful in modelling the hydrogel efficiency of water-absorbing additives. The paper presents the results of a study of the physical and chemical properties of hydrogels based on of the production of "Nitron" (Polyacrylonitrile) wastes fibre and salts of the 3-rd transition metals and formalin. The developed hydrogels HG-Al, HG-Cr and HG-formalin have been tested for water holding the capacity of sand. Such a conclusion was also confirmed by data from the method of determining the wilting point by vegetative thumbnails. In the entering process using a dose of 0.1% of the swelling polymeric hydrogel in sand with a culture of barley the difference between the wilting point in comparison with the control was negligible. This indicates that the moisture which was contained in the hydrogel is involved in moisture availability for plant growth, to the same extent as that in the capillaries.Keywords: hydrogel, chemical, polymer, sandy, colloid
Procedia PDF Downloads 14827175 Secure Data Sharing of Electronic Health Records With Blockchain
Authors: Kenneth Harper
Abstract:
The secure sharing of Electronic Health Records (EHRs) is a critical challenge in modern healthcare, demanding solutions to enhance interoperability, privacy, and data integrity. Traditional standards like Health Information Exchange (HIE) and HL7 have made significant strides in facilitating data exchange between healthcare entities. However, these approaches rely on centralized architectures that are often vulnerable to data breaches, lack sufficient privacy measures, and have scalability issues. This paper proposes a framework for secure, decentralized sharing of EHRs using blockchain technology, cryptographic tokens, and Non-Fungible Tokens (NFTs). The blockchain's immutable ledger, decentralized control, and inherent security mechanisms are leveraged to improve transparency, accountability, and auditability in healthcare data exchanges. Furthermore, we introduce the concept of tokenizing patient data through NFTs, creating unique digital identifiers for each record, which allows for granular data access controls and proof of data ownership. These NFTs can also be employed to grant access to authorized parties, establishing a secure and transparent data sharing model that empowers both healthcare providers and patients. The proposed approach addresses common privacy concerns by employing privacy-preserving techniques such as zero-knowledge proofs (ZKPs) and homomorphic encryption to ensure that sensitive patient information can be shared without exposing the actual content of the data. This ensures compliance with regulations like HIPAA and GDPR. Additionally, the integration of Fast Healthcare Interoperability Resources (FHIR) with blockchain technology allows for enhanced interoperability, enabling healthcare organizations to exchange data seamlessly and securely across various systems while maintaining data governance and regulatory compliance. Through real-world case studies and simulations, this paper demonstrates how blockchain-based EHR sharing can reduce operational costs, improve patient outcomes, and enhance the security and privacy of healthcare data. This decentralized framework holds great potential for revolutionizing healthcare information exchange, providing a transparent, scalable, and secure method for managing patient data in a highly regulated environment.Keywords: blockchain, electronic health records (ehrs), fast healthcare interoperability resources (fhir), health information exchange (hie), hl7, interoperability, non-fungible tokens (nfts), privacy-preserving techniques, tokens, secure data sharing,
Procedia PDF Downloads 2727174 The Applications of Four Fingers Theory: The Proof of 66 Acupoints under the Human Elbow and Knee
Authors: Chih-I. Tsai, Yu-Chien. Lin
Abstract:
Through experiences of clinical practices, it is discovered that locations on the body at a level of four fingerbreadth above and below the joints are the points at which muscles connect to tendons, and since the muscles and tendons possess opposite characteristics, muscles are full of blood but lack qi, while tendons are full of qi but lack blood, these points on our body become easily blocked. It is proposed that through doing acupuncture or creating localized pressure to the areas four fingerbreadths above and below our joints, with an elastic bandage, we could help the energy, also known as qi, to flow smoothly in our body and further improve our health. Based on the Four Fingers Theory, we understand that human height is 22 four fingerbreadths. In addition, qi and blood travel through 24 meridians, 50 times each day, and they flow through 6 cun with every human breath. We can also understand the average number of human heartbeats is 75 times per minute. And the function of qi-blood circulation system in Traditional Chinese Medicine is the same as the blood circulation in Western Medical Science. Informed by Four Fingers Theory, this study further examined its applications in acupuncture practices. The research question is how Four Fingers Theory proves what has been mentioned in Nei Jing that there are 66 acupoints under a human’s elbow and knee. In responding to the research question, there are 66 acupoints under a human’s elbow and knee. Four Fingers Theory facilitated the creation of the acupuncture naming and teaching system. It is expected to serve as an approachable and effective way to deliver knowledge of acupuncture to the public worldwide.Keywords: four fingers theory, meridians circulation, 66 acupoints under human elbow and knee, acupuncture
Procedia PDF Downloads 30327173 The Twin Terminal of Pedestrian Trajectory Based on City Intelligent Model (CIM) 4.0
Authors: Chen Xi, Liu Xuebing, Lao Xueru, Kuan Sinman, Jiang Yike, Wang Hanwei, Yang Xiaolang, Zhou Junjie, Xie Jinpeng
Abstract:
To further promote the development of smart cities, the microscopic "nerve endings" of the City Intelligent Model (CIM) are extended to be more sensitive. In this paper, we develop a pedestrian trajectory twin terminal based on the CIM and CNN technology. It also uses 5G networks, architectural and geoinformatics technologies, convolutional neural networks, combined with deep learning networks for human behavior recognition models, to provide empirical data such as 'pedestrian flow data and human behavioral characteristics data', and ultimately form spatial performance evaluation criteria and spatial performance warning systems, to make the empirical data accurate and intelligent for prediction and decision making.Keywords: urban planning, urban governance, CIM, artificial intelligence, sustainable development
Procedia PDF Downloads 42627172 Potential Determinants of Research Output: Comparing Economics and Business
Authors: Osiris Jorge Parcero, Néstor Gandelman, Flavia Roldán, Josef Montag
Abstract:
This paper uses cross-country unbalanced panel data of up to 146 countries over the period 1996 to 2015 to be the first study to identify potential determinants of a country’s relative research output in Economics versus Business. More generally, it is also one of the first studies comparing Economics and Business. The results show that better policy-related data availability, higher income inequality, and lower ethnic fractionalization relatively favor economics. The findings are robust to two alternative fixed effects specifications, three alternative definitions of economics and business, two alternative measures of research output (publications and citations), and the inclusion of meaningful control variables. To the best of our knowledge, our paper is also the first to demonstrate the importance of policy-related data as drivers of economic research. Our regressions show that the availability of this type of data is the single most important factor associated with the prevalence of economics over business as a research domain. Thus, our work has policy implications, as the availability of policy-related data is partially under policy control. Moreover, it has implications for students, professionals, universities, university departments, and research-funding agencies that face choices between profiles oriented toward economics and those oriented toward business. Finally, the conclusions show potential lines for further research.Keywords: research output, publication performance, bibliometrics, economics, business, policy-related data
Procedia PDF Downloads 13827171 Assessment of Routine Health Information System (RHIS) Quality Assurance Practices in Tarkwa Sub-Municipal Health Directorate, Ghana
Authors: Richard Okyere Boadu, Judith Obiri-Yeboah, Kwame Adu Okyere Boadu, Nathan Kumasenu Mensah, Grace Amoh-Agyei
Abstract:
Routine health information system (RHIS) quality assurance has become an important issue, not only because of its significance in promoting a high standard of patient care but also because of its impact on government budgets for the maintenance of health services. A routine health information system comprises healthcare data collection, compilation, storage, analysis, report generation, and dissemination on a routine basis in various healthcare settings. The data from RHIS give a representation of health status, health services, and health resources. The sources of RHIS data are normally individual health records, records of services delivered, and records of health resources. Using reliable information from routine health information systems is fundamental in the healthcare delivery system. Quality assurance practices are measures that are put in place to ensure the health data that are collected meet required quality standards. Routine health information system quality assurance practices ensure that data that are generated from the system are fit for use. This study considered quality assurance practices in the RHIS processes. Methods: A cross-sectional study was conducted in eight health facilities in Tarkwa Sub-Municipal Health Service in the western region of Ghana. The study involved routine quality assurance practices among the 90 health staff and management selected from facilities in Tarkwa Sub-Municipal who collected or used data routinely from 24th December 2019 to 20th January 2020. Results: Generally, Tarkwa Sub-Municipal health service appears to practice quality assurance during data collection, compilation, storage, analysis and dissemination. The results show some achievement in quality control performance in report dissemination (77.6%), data analysis (68.0%), data compilation (67.4%), report compilation (66.3%), data storage (66.3%) and collection (61.1%). Conclusions: Even though the Tarkwa Sub-Municipal Health Directorate engages in some control measures to ensure data quality, there is a need to strengthen the process to achieve the targeted percentage of performance (90.0%). There was a significant shortfall in quality assurance practices performance, especially during data collection, with respect to the expected performance.Keywords: quality assurance practices, assessment of routine health information system quality, routine health information system, data quality
Procedia PDF Downloads 8627170 Critically Analyzing the Application of Big Data for Smart Transportation: A Case Study of Mumbai
Authors: Tanuj Joshi
Abstract:
Smart transportation is fast emerging as a solution to modern cities’ approach mobility issues, delayed emergency response rate and high congestion on streets. Present day scenario with Google Maps, Waze, Yelp etc. demonstrates how information and communications technologies controls the intelligent transportation system. This intangible and invisible infrastructure is largely guided by the big data analytics. On the other side, the exponential increase in Indian urban population has intensified the demand for better services and infrastructure to satisfy the transportation needs of its citizens. No doubt, India’s huge internet usage is looked as an important resource to guide to achieve this. However, with a projected number of over 40 billion objects connected to the Internet by 2025, the need for systems to handle massive volume of data (big data) also arises. This research paper attempts to identify the ways of exploiting the big data variables which will aid commuters on Indian tracks. This study explores real life inputs by conducting survey and interviews to identify which gaps need to be targeted to better satisfy the customers. Several experts at Mumbai Metropolitan Region Development Authority (MMRDA), Mumbai Metro and Brihanmumbai Electric Supply and Transport (BEST) were interviewed regarding the Information Technology (IT) systems currently in use. The interviews give relevant insights and requirements into the workings of public transportation systems whereas the survey investigates the macro situation.Keywords: smart transportation, mobility issue, Mumbai transportation, big data, data analysis
Procedia PDF Downloads 18127169 Corrosion Behavior of Organic-Inorganic Hybrid Coatings Fabricated by Electrostatic Method
Authors: Mohammed Ahmed, Ziba Nazarlou
Abstract:
Mild steels have a limited alloying content which makes them vulnerable to excessive corrosion rates in the harsh medium. To overcome this issue, some protective coatings are used to prevent corrosion on the steel surface. The use of specialized coatings, mainly organic coatings (such as epoxies, polyurethanes, and acrylics) and inorganic coatings (such as Polysiloxanes) is the most common method of mitigating corrosion of carbon steel. Incorporating the benefits of organic and inorganic hybrid (OIH) compounds for the designing of hybrid protective coatings is still challenging for industrial applications. There are advantages of inorganic coatings have, but purely inorganic siloxane-based coatings are difficult to use on industrial applications unless they are used at extremely low thicknesses (< 1-2 microns). Hence, most industrial applications try to have a combination of Polysiloxanes with organic compounds. A hybrid coating possesses an organic section, which transports flexibility and impact resistance, and an inorganic section, which usually helps in the decreasing of porosity and increasing thermal stability and hardness. A number of polymers including polyethylene glycol and polyvinyl pyrrolidone have been reported to inhibit the corrosion mild steel in acidic media. However, reports on the effect of polyethylene oxide (PEO) or its blends on corrosion inhibition of metals is very scarce. Different composition of OIH coatings was synthesized by using silica sol-gel, epoxy, and PEO. The effect of different coating types on the corrosion behavior of carbon steel in harsh solution has been studied by weight loss and electrochemical measurements using Gamry 1000 Interface Potentiostat. Coating structures were investigated by SEM. İt revealed a considerable reduction in corrosion rate for coated sample. Based on these results, OIH coating prepared by epoxy-silica sol gel-PEO and epoxy-silica sol-gel exhibit had a %99.5 and %98 reduction of (Corrosion rate) CR compares to baseline. Cathodic Tafel constant (βc) shows that coatings change both Tafel constants but had more effect on the cathodic process. The evolution of the Potentiostatic scan with time displays stability in potential, some of them in a high value while the other in a low value which can be attributed to the formation of an oxide film covering substrate surface. The coated samples with the group of epoxy coating have a lower potential along with the time test, while the silica group shows higher in potential with respect to time.Keywords: electrostatic, hybrid coating, corrosion tests, silica sol gel
Procedia PDF Downloads 12327168 Aircraft Automatic Collision Avoidance Using Spiral Geometric Approach
Authors: M. Orefice, V. Di Vito
Abstract:
This paper provides a description of a Collision Avoidance algorithm that has been developed starting from the mathematical modeling of the flight of insects, in terms of spirals and conchospirals geometric paths. It is able to calculate a proper avoidance manoeuver aimed to prevent the infringement of a predefined distance threshold between ownship and the considered intruder, while minimizing the ownship trajectory deviation from the original path and in compliance with the aircraft performance limitations and dynamic constraints. The algorithm is designed in order to be suitable for real-time applications, so that it can be considered for the implementation in the most recent airborne automatic collision avoidance systems using the traffic data received through an ADS-B IN device. The presented approach is able to take into account the rules-of-the-air, due to the possibility to select, through specifically designed decision making logic based on the consideration of the encounter geometry, the direction of the calculated collision avoidance manoeuver that allows complying with the rules-of-the-air, as for instance the fundamental right of way rule. In the paper, the proposed collision avoidance algorithm is presented and its preliminary design and software implementation is described. The applicability of this method has been proved through preliminary simulation tests performed in a 2D environment considering single intruder encounter geometries, as reported and discussed in the paper.Keywords: ADS-B Based Application, Collision Avoidance, RPAS, Spiral Geometry.
Procedia PDF Downloads 24427167 Investigating the Effects of Data Transformations on a Bi-Dimensional Chi-Square Test
Authors: Alexandru George Vaduva, Adriana Vlad, Bogdan Badea
Abstract:
In this research, we conduct a Monte Carlo analysis on a two-dimensional χ2 test, which is used to determine the minimum distance required for independent sampling in the context of chaotic signals. We investigate the impact of transforming initial data sets from any probability distribution to new signals with a uniform distribution using the Spearman rank correlation on the χ2 test. This transformation removes the randomness of the data pairs, and as a result, the observed distribution of χ2 test values differs from the expected distribution. We propose a solution to this problem and evaluate it using another chaotic signal.Keywords: chaotic signals, logistic map, Pearson’s test, Chi Square test, bivariate distribution, statistical independence
Procedia PDF Downloads 10227166 Real Time Data Communication with FlightGear Using Simulink Over a UDP Protocol
Authors: Adil Loya, Ali Haider, Arslan A. Ghaffor, Abubaker Siddique
Abstract:
Simulation and modelling of Unmanned Aero Vehicle (UAV) has gained wide popularity in front of aerospace community. The demand of designing and modelling optimized control system for UAV has increased ten folds since last decade. The reason is next generation warfare is dependent on unmanned technologies. Therefore, this research focuses on the simulation of nonlinear UAV dynamics on Simulink and its integration with Flightgear. There has been lots of research on implementation of optimizing control using Simulink, however, there are fewer known techniques to simulate these dynamics over Flightgear and a tedious technique of acquiring data has been tackled in this research horizon. Sending data to Flightgear is easy but receiving it from Simulink is not that straight forward, i.e. we can only receive control data on the output. However, in this research we have managed to get the data out from the Flightgear by implementation of level 2 s-function block within Simulink. Moreover, the results captured from Flightgear over a Universal Datagram Protocol (UDP) communication are then compared with the attitude signal that were sent previously. This provide useful information regarding the difference in outputs attained from Simulink to Flightgear. It was found that values received on Simulink were in high agreement with that of the Flightgear output. And complete study has been conducted in a discrete way.Keywords: aerospace, flight control, flightgear, communication, Simulink
Procedia PDF Downloads 29127165 Product Features Extraction from Opinions According to Time
Authors: Kamal Amarouche, Houda Benbrahim, Ismail Kassou
Abstract:
Nowadays, e-commerce shopping websites have experienced noticeable growth. These websites have gained consumers’ trust. After purchasing a product, many consumers share comments where opinions are usually embedded about the given product. Research on the automatic management of opinions that gives suggestions to potential consumers and portrays an image of the product to manufactures has been growing recently. After launching the product in the market, the reviews generated around it do not usually contain helpful information or generic opinions about this product (e.g. telephone: great phone...); in the sense that the product is still in the launching phase in the market. Within time, the product becomes old. Therefore, consumers perceive the advantages/ disadvantages about each specific product feature. Therefore, they will generate comments that contain their sentiments about these features. In this paper, we present an unsupervised method to extract different product features hidden in the opinions which influence its purchase, and that combines Time Weighting (TW) which depends on the time opinions were expressed with Term Frequency-Inverse Document Frequency (TF-IDF). We conduct several experiments using two different datasets about cell phones and hotels. The results show the effectiveness of our automatic feature extraction, as well as its domain independent characteristic.Keywords: opinion mining, product feature extraction, sentiment analysis, SentiWordNet
Procedia PDF Downloads 41727164 Prediction of Gully Erosion with Stochastic Modeling by using Geographic Information System and Remote Sensing Data in North of Iran
Authors: Reza Zakerinejad
Abstract:
Gully erosion is a serious problem that threading the sustainability of agricultural area and rangeland and water in a large part of Iran. This type of water erosion is the main source of sedimentation in many catchment areas in the north of Iran. Since in many national assessment approaches just qualitative models were applied the aim of this study is to predict the spatial distribution of gully erosion processes by means of detail terrain analysis and GIS -based logistic regression in the loess deposition in a case study in the Golestan Province. This study the DEM with 25 meter result ion from ASTER data has been used. The Landsat ETM data have been used to mapping of land use. The TreeNet model as a stochastic modeling was applied to prediction the susceptible area for gully erosion. In this model ROC we have set 20 % of data as learning and 20 % as learning data. Therefore, applying the GIS and satellite image analysis techniques has been used to derive the input information for these stochastic models. The result of this study showed a high accurate map of potential for gully erosion.Keywords: TreeNet model, terrain analysis, Golestan Province, Iran
Procedia PDF Downloads 54127163 A Kinetic Study of Radical Polymerisation of Acrylic Monomers in the Presence of the Liquid Crystal and the Electro-Optical Properties of These Mixtures
Authors: A. Bouriche, D. Merah, T. Bouchaour, L. Alachaher-Bedjaoui, U. Maschke
Abstract:
Intensive research continues in the field of liquid crystals (LCs) for their potential use in modern display applications. Nematic LCs has been most commonly used due to the large birefringence and their sensitivity to even weak perturbation forces induced by electric, magnetic and optical fields. Polymer dispersed liquid crystals (PDLCs), composed of micron-sized nematic LC droplets dispersed in a polymer matrix is an important class of materials for applications in different domains of technology involving large area display devices, optical switches, phase modulators, variable attenuators, polarisers, flexible displays and smart windows. In this study the composites are prepared from mixtures of mono functional acrylic monomers, (Butylacrylate (ABu), 2-Ethylhexylacrylate (2-EHA), 2-Hydroxyethyl methacrylate (HEMA) and hydroxybutylmethacrylate (HBMA)) and two liquid crystals: (4-cyano-4'-n-pentyl-biphenyl) (5CB) and E7 which is an eutectic mixtures of four cyanoparaphenylenes. These mixtures are prepared adding the Darocur 1173 as photoinitiator, the 1.6-hexanediol diacrylate (HDDA) as cross-linker agent, and finally they are exposed to UV irradiation. The kinetic polymerization of monomer/LC mixture were investigated with the Fourier Transform Infra Red spectroscopy (FTIR). The electro-optical properties of the PDLC films were determined by measuring the voltage dependence on the transmitted light.Keywords: acrylic monomers, films PDLC, liquid crystal, polymerisation
Procedia PDF Downloads 29627162 Data Science/Artificial Intelligence: A Possible Panacea for Refugee Crisis
Authors: Avi Shrivastava
Abstract:
In 2021, two heart-wrenching scenes, shown live on television screens across countries, painted a grim picture of refugees. One of them was of people clinging onto an airplane's wings in their desperate attempt to flee war-torn Afghanistan. They ultimately fell to their death. The other scene was the U.S. government authorities separating children from their parents or guardians to deter migrants/refugees from coming to the U.S. These events show the desperation refugees feel when they are trying to leave their homes in disaster zones. However, data paints a grave picture of the current refugee situation. It also indicates that a bleak future lies ahead for the refugees across the globe. Data and information are the two threads that intertwine to weave the shimmery fabric of modern society. Data and information are often used interchangeably, but they differ considerably. For example, information analysis reveals rationale, and logic, while data analysis, on the other hand, reveals a pattern. Moreover, patterns revealed by data can enable us to create the necessary tools to combat huge problems on our hands. Data analysis paints a clear picture so that the decision-making process becomes simple. Geopolitical and economic data can be used to predict future refugee hotspots. Accurately predicting the next refugee hotspots will allow governments and relief agencies to prepare better for future refugee crises. The refugee crisis does not have binary answers. Given the emotionally wrenching nature of the ground realities, experts often shy away from realistically stating things as they are. This hesitancy can cost lives. When decisions are based solely on data, emotions can be removed from the decision-making process. Data also presents irrefutable evidence and tells whether there is a solution or not. Moreover, it also responds to a nonbinary crisis with a binary answer. Because of all that, it becomes easier to tackle a problem. Data science and A.I. can predict future refugee crises. With the recent explosion of data due to the rise of social media platforms, data and insight into data has solved many social and political problems. Data science can also help solve many issues refugees face while staying in refugee camps or adopted countries. This paper looks into various ways data science can help solve refugee problems. A.I.-based chatbots can help refugees seek legal help to find asylum in the country they want to settle in. These chatbots can help them find a marketplace where they can find help from the people willing to help. Data science and technology can also help solve refugees' many problems, including food, shelter, employment, security, and assimilation. The refugee problem seems to be one of the most challenging for social and political reasons. Data science and machine learning can help prevent the refugee crisis and solve or alleviate some of the problems that refugees face in their journey to a better life. With the explosion of data in the last decade, data science has made it possible to solve many geopolitical and social issues.Keywords: refugee crisis, artificial intelligence, data science, refugee camps, Afghanistan, Ukraine
Procedia PDF Downloads 7527161 Scheduling of Cross-Docking Center: An Auction-Based Algorithm
Authors: Eldho Paul, Brijesh Paul
Abstract:
This work proposes an auction mechanism based solution methodology for the optimum scheduling of trucks in a cross-docking centre. The cross-docking centre is an important element of lean supply chain. It reduces the amount of storage and transportation costs in the distribution system compared to an ordinary warehouse. Better scheduling of trucks in a cross-docking center is the best way to reduce storage and transportation costs. Auction mechanism is commonly used for allocation of limited resources in different real-life applications. Here, we try to schedule inbound trucks by integrating auction mechanism with the functioning of a cross-docking centre. A mathematical model is developed for the optimal scheduling of inbound trucks based on the auction methodology. The determination of exact solution for problems involving large number of trucks was found to be computationally difficult, and hence a genetic algorithm based heuristic methodology is proposed in this work. A comparative study of exact and heuristic solutions is done using five classes of data sets. It is observed from the study that the auction-based mechanism is capable of providing good solutions to scheduling problem in cross-docking centres.Keywords: auction mechanism, cross-docking centre, genetic algorithm, scheduling of trucks
Procedia PDF Downloads 41627160 Image Segmentation Techniques: Review
Authors: Lindani Mbatha, Suvendi Rimer, Mpho Gololo
Abstract:
Image segmentation is the process of dividing an image into several sections, such as the object's background and the foreground. It is a critical technique in both image-processing tasks and computer vision. Most of the image segmentation algorithms have been developed for gray-scale images and little research and algorithms have been developed for the color images. Most image segmentation algorithms or techniques vary based on the input data and the application. Nearly all of the techniques are not suitable for noisy environments. Most of the work that has been done uses the Markov Random Field (MRF), which involves the computations and is said to be robust to noise. In the past recent years' image segmentation has been brought to tackle problems such as easy processing of an image, interpretation of the contents of an image, and easy analysing of an image. This article reviews and summarizes some of the image segmentation techniques and algorithms that have been developed in the past years. The techniques include neural networks (CNN), edge-based techniques, region growing, clustering, and thresholding techniques and so on. The advantages and disadvantages of medical ultrasound image segmentation techniques are also discussed. The article also addresses the applications and potential future developments that can be done around image segmentation. This review article concludes with the fact that no technique is perfectly suitable for the segmentation of all different types of images, but the use of hybrid techniques yields more accurate and efficient results.Keywords: clustering-based, convolution-network, edge-based, region-growing
Procedia PDF Downloads 10227159 Anti-Inflammatory and Analgesic Effects of Methanol Extract of Rhizophora racemosa Leaf in Albino Rats
Authors: Angalabiri-Owei E. Bekekeme, Brambaifa Nelson
Abstract:
In view of the peculiar environment of the Niger Delta, access to modern health care is limited, hence the inhabitants especially those in the swampy areas resorts to sourcing for alternatives cure for their ailments using plants commonly found in this area without scientific evaluation. Rhizophora racemosa, G. F. Meyer (Rhizophoraceae) is the most abundant mangrove plant in the Niger Delta Area of Nigeria. The plant has been observed to be used for relief of a toothache and dysmenorrhoea among some Ijaw communities in the region. This work has revealed the likely potential of the plant in drug discovery and development. The crude methanol extract at doses of 300 mg/kg and 600 mg/kg (intraperitoneal) were tested for analgesic effect using fresh egg albumin induced inflammatory pain and Randall–Sellito method to assess the pain threshold. The anti-inflammatory effect was also evaluated with the extract at doses of 300 mg/kg and 600 mg/kg (intraperitoneal) using acute inflammatory model; fresh egg albumin induced paw oedema and assessed using Plethysmometer in rats. The methanol extracts 300 mg/kg and 600 mg/kg exhibited a significant (P < 0.001) and dose-dependent analgesic activity compared with the negative control and a standard drug diclofenac using ANOVA with Least Significant Difference post hoc test as evidenced by increased pain threshold. Also, the extract significantly (P < 0.001) reduced the rat paw oedema induced by the sub plantar injection of fresh egg albumin when compared with the negative control and a standard diclofenac using above statistical methods. This study revealed that the plant possesses analgesic and anti-inflammatory activities hence provide scientific bases for use as medicine.Keywords: analgesic, anti-inflammatory, plethysmometer, Rhizophora racemosa
Procedia PDF Downloads 36327158 Aza-Flavanones as Small Molecule Inhibitors of MicroRNA-10b in MDA-MB-231 Breast Cancer Cells
Authors: Debasmita Mukhopadhyay, Manika Pal Bhadra
Abstract:
MiRNAs contribute to oncogenesis either as tumor suppressors or oncogenes. Hence, discovery of miRNA-based therapeutics are imperative to ameliorate cancer. Modulation of miRNA maturation is accomplished via several therapeutic agents, including small molecules and oligonucleotides. Due to the attractive pharmacokinetic properties of small molecules over oligonucleotides, we set to identify small molecule inhibitors of a metastasis-inducing microRNA. Cytotoxicity profile of aza-flavanone C1 was analyzed in a panel of breast cancer cells employing the NCI-60 screen protocols. Flow cytometry, immunofluorescence and western blotting of apoptotic or EMT markers were performed to analyze the effect of C1. A dual luciferase assay unequivocally suggested that C1 repressed endogenous miR-10b in MDA-MB-231 cells. A derivative of aza-flavanone C1 is shown as a strong inhibitor miR-10b. Blockade of miR-10b by C1 resulted in decreased expression of miR-10b targets in an aggressive breast cancer cell line model, MDA-MB-231. Abrogation of TWIST1, an EMT-inducing transcription factor also contributed to C1 mediated apoptosis. Moreover C1 exhibited a specific and selective down-regulation of miR-10b and did not function as a general inhibitor of miRNA biogenesis or other oncomiRs of breast carcinoma. Aza-flavanone congener C1 functions as a potent inhibitor of the metastasis-inducing microRNA, miR-10b. Our present study provides evidence for targeting metastasis-inducing microRNA, miR-10b with a derivative of Aza-flavanone. Better pharmacokinetic properties of small molecules place them as attractive agents compared to nucleic acids based therapies to target miRNA. Further work, in generating analogues based on aza-flavanone moieties will significantly improve the affinity of the small molecules to bind miR-10b. Finally, it is imperative to develop small molecules as novel miRNA-therapeutics in the fight against cancer.Keywords: breast cancer, microRNA, metastasis, EMT
Procedia PDF Downloads 56927157 A Literature Review of Precision Agriculture: Applications of Diagnostic Diseases in Corn, Potato, and Rice Based on Artificial Intelligence
Authors: Carolina Zambrana, Grover Zurita
Abstract:
The food loss production that occurs in deficient agricultural production is one of the major problems worldwide. This puts the population's food security and the efficiency of farming investments at risk. It is to be expected that this food security will be achieved with the own and efficient production of each country. It will have an impact on the well-being of its population and, thus, also on food sovereignty. The production losses in quantity and quality occur due to the lack of efficient detection of diseases at an early stage. It is very difficult to solve the agriculture efficiency using traditional methods since it takes a long time to be carried out due to detection imprecision of the main diseases, especially when the production areas are extensive. Therefore, the main objective of this research study is to perform a systematic literature review, of the latest five years, of Precision Agriculture (PA) to be able to understand the state of the art of the set of new technologies, procedures, and optimization processes with Artificial Intelligence (AI). This study will focus on Corns, Potatoes, and Rice diagnostic diseases. The extensive literature review will be performed on Elsevier, Scopus, and IEEE databases. In addition, this research will focus on advanced digital imaging processing and the development of software and hardware for PA. The convolution neural network will be handling special attention due to its outstanding diagnostic results. Moreover, the studied data will be incorporated with artificial intelligence algorithms for the automatic diagnosis of crop quality. Finally, precision agriculture with technology applied to the agricultural sector allows the land to be exploited efficiently. This system requires sensors, drones, data acquisition cards, and global positioning systems. This research seeks to merge different areas of science, control engineering, electronics, digital image processing, and artificial intelligence for the development, in the near future, of a low-cost image measurement system that allows the optimization of crops with AI.Keywords: precision agriculture, convolutional neural network, deep learning, artificial intelligence
Procedia PDF Downloads 8527156 Exploring Reading into Writing: A Corpus-Based Analysis of Postgraduate Students’ Literature Review Essays
Authors: Tanzeela Anbreen, Ammara Maqsood
Abstract:
Reading into writing is one of university students' most required academic skills. The current study explored postgraduate university students’ writing quality using a corpus-based approach. Twelve postgraduate students’ literature review essays were chosen for the corpus-based analysis. These essays were chosen because students had to incorporate multiple reading sources in these essays, which was a new writing exercise for them. The students were provided feedback at least two times which comprised of the written comments by the tutor highlighting the areas of improvement and also by using the ‘track changes’ function. This exercise was repeated two times, and students submitted two drafts. This investigation included only the finally submitted work of the students. A corpus-based approach was adopted to analyse the essays because it promotes autonomous discovery and personalised learning. The aim of this analysis was to understand the existing level of students’ writing before the start of their postgraduate thesis. Text Inspector was used to analyse the quality of essays. With the help of the Text Inspector tool, the vocabulary used in the essays was compared to the English Vocabulary Profile (EVP), which describes what learners know and can do at each Common European Framework of Reference (CEFR) level. Writing quality was also measured for the Flesch reading ease score, which is a standard to describe the ease of understanding the writing content. The results reflected that students found writing essays using multiple sources challenging. In most essays, the vocabulary level achieved was between B1-B2 of the CEFL level. The study recommends that students need extensive training in developing academic writing skills, particularly in writing the literature review type assignment, which requires multiple sources citations.Keywords: literature review essays, postgraduate students, corpus-based analysis, vocabulary proficiency
Procedia PDF Downloads 7727155 Bayesian Variable Selection in Quantile Regression with Application to the Health and Retirement Study
Authors: Priya Kedia, Kiranmoy Das
Abstract:
There is a rich literature on variable selection in regression setting. However, most of these methods assume normality for the response variable under consideration for implementing the methodology and establishing the statistical properties of the estimates. In many real applications, the distribution for the response variable may be non-Gaussian, and one might be interested in finding the best subset of covariates at some predetermined quantile level. We develop dynamic Bayesian approach for variable selection in quantile regression framework. We use a zero-inflated mixture prior for the regression coefficients, and consider the asymmetric Laplace distribution for the response variable for modeling different quantiles of its distribution. An efficient Gibbs sampler is developed for our computation. Our proposed approach is assessed through extensive simulation studies, and real application of the proposed approach is also illustrated. We consider the data from health and retirement study conducted by the University of Michigan, and select the important predictors when the outcome of interest is out-of-pocket medical cost, which is considered as an important measure for financial risk. Our analysis finds important predictors at different quantiles of the outcome, and thus enhance our understanding on the effects of different predictors on the out-of-pocket medical cost.Keywords: variable selection, quantile regression, Gibbs sampler, asymmetric Laplace distribution
Procedia PDF Downloads 159