Search results for: ground data
25583 Spatial Distribution and Habitat Preference of Indian Pangolin (Manis crassicaudata) in Madhesh Province, Nepal
Authors: Asmit Neupane, Narayan Prasad Gautam, Prabin Bhusal
Abstract:
Indian pangolin, locally called as ‘Salak’, ‘Sal machha’, ‘Pakho machha’, is a globally endangered species, nationally categorized as a critically endangered species, protected under the National Parks and Wildlife Conservation (NPWC) Act 1973 and appended in Appendix I of CITES. Indian pangolins occur in the tropical areas of Terai region and Chure foothills of eastern Nepal, and India, Bangladesh, Pakistan, and Sri Lanka. They utilize a wide range of habitats, including primary and secondary tropical forest, limestone forest, bamboo forest, grassland, and agricultural lands. So, in regard to this fact, this research is aimed to provide detailed information regarding the current distribution pattern, status, habitat preference, prevailing threats and attitude of local people towards species conservation in Madhesh Province, Nepal. The study was conducted in four CFs, two from Bara district and two from Dhanusha district. The study area comprised of Churia range and foothills with tropical and sub-tropical vegetation. A total of 24 transects were established, each of 500*50 m2, where indirect signs of Indian pangolin, including active/old burrows, pugmarks and scratches, were found. Altogether 93 burrows were found, where only 20 were active burrows. Similarly, a vegetation survey and social survey was also conducted. The data was analyzed using Stata 16 and SPSS software. Distance from settlement, ground cover, aspect, presence/absence of ants/termites and human disturbance were the important habitat parameters having statistically significant relationship with the distribution of Indian pangolin in the area. The species was found to prefer an elevation of 360 to 540m, 0-15º slope, red soil, North-east aspect, moderate crown and ground cover, without fire and rocks, vicinity of water, roads, settlement, Sal dominated forest and minimum disturbed by human activities. Similarly, the attitude of local people towards Indian pangolin conservation was found to be significantly different with respect to age, sex and education level. The study concludes that majority of active burrows were found in Churia hills, which indicates that Indian pangolin population is gradually moving uphill towards higher elevation as hilly area supports better prey availability and also less human disturbance. Further studies are required to investigate microhabitat preferences, seasonal variability and impacts of climate change on the distribution, habitat and prey availability of Indian pangolin for the sustainable conservation of this species.Keywords: conservation, IUCN red list, local participation, small mammal, status, threats
Procedia PDF Downloads 8025582 Efficient Schemes of Classifiers for Remote Sensing Satellite Imageries of Land Use Pattern Classifications
Authors: S. S. Patil, Sachidanand Kini
Abstract:
Classification of land use patterns is compelling in complexity and variability of remote sensing imageries data. An imperative research in remote sensing application exploited to mine some of the significant spatially variable factors as land cover and land use from satellite images for remote arid areas in Karnataka State, India. The diverse classification techniques, unsupervised and supervised consisting of maximum likelihood, Mahalanobis distance, and minimum distance are applied in Bellary District in Karnataka State, India for the classification of the raw satellite images. The accuracy evaluations of results are compared visually with the standard maps with ground-truths. We initiated with the maximum likelihood technique that gave the finest results and both minimum distance and Mahalanobis distance methods over valued agriculture land areas. In meanness of mislaid few irrelevant features due to the low resolution of the satellite images, high-quality accord between parameters extracted automatically from the developed maps and field observations was found.Keywords: Mahalanobis distance, minimum distance, supervised, unsupervised, user classification accuracy, producer's classification accuracy, maximum likelihood, kappa coefficient
Procedia PDF Downloads 18325581 Seismic Response Control of Multi-Span Bridge Using Magnetorheological Dampers
Authors: B. Neethu, Diptesh Das
Abstract:
The present study investigates the performance of a semi-active controller using magneto-rheological dampers (MR) for seismic response reduction of a multi-span bridge. The application of structural control to the structures during earthquake excitation involves numerous challenges such as proper formulation and selection of the control strategy, mathematical modeling of the system, uncertainty in system parameters and noisy measurements. These problems, however, need to be tackled in order to design and develop controllers which will efficiently perform in such complex systems. A control algorithm, which can accommodate un-certainty and imprecision compared to all the other algorithms mentioned so far, due to its inherent robustness and ability to cope with the parameter uncertainties and imprecisions, is the sliding mode algorithm. A sliding mode control algorithm is adopted in the present study due to its inherent stability and distinguished robustness to system parameter variation and external disturbances. In general a semi-active control scheme using an MR damper requires two nested controllers: (i) an overall system controller, which derives the control force required to be applied to the structure and (ii) an MR damper voltage controller which determines the voltage required to be supplied to the damper in order to generate the desired control force. In the present study a sliding mode algorithm is used to determine the desired optimal force. The function of the voltage controller is to command the damper to produce the desired force. The clipped optimal algorithm is used to find the command voltage supplied to the MR damper which is regulated by a semi active control law based on sliding mode algorithm. The main objective of the study is to propose a robust semi active control which can effectively control the responses of the bridge under real earthquake ground motions. Lumped mass model of the bridge is developed and time history analysis is carried out by solving the governing equations of motion in the state space form. The effectiveness of MR dampers is studied by analytical simulations by subjecting the bridge to real earthquake records. In this regard, it may also be noted that the performance of controllers depends, to a great extent, on the characteristics of the input ground motions. Therefore, in order to study the robustness of the controller in the present study, the performance of the controllers have been investigated for fourteen different earthquake ground motion records. The earthquakes are chosen in such a way that all possible characteristic variations can be accommodated. Out of these fourteen earthquakes, seven are near-field and seven are far-field. Also, these earthquakes are divided into different frequency contents, viz, low-frequency, medium-frequency, and high-frequency earthquakes. The responses of the controlled bridge are compared with the responses of the corresponding uncontrolled bridge (i.e., the bridge without any control devices). The results of the numerical study show that the sliding mode based semi-active control strategy can substantially reduce the seismic responses of the bridge showing a stable and robust performance for all the earthquakes.Keywords: bridge, semi active control, sliding mode control, MR damper
Procedia PDF Downloads 12425580 Field Trial of Resin-Based Composite Materials for the Treatment of Surface Collapses Associated with Former Shallow Coal Mining
Authors: Philip T. Broughton, Mark P. Bettney, Isla L. Smail
Abstract:
Effective treatment of ground instability is essential when managing the impacts associated with historic mining. A field trial was undertaken by the Coal Authority to investigate the geotechnical performance and potential use of composite materials comprising resin and fill or stone to safely treat surface collapses, such as crown-holes, associated with shallow mining. Test pits were loosely filled with various granular fill materials. The fill material was injected with commercially available silicate and polyurethane resin foam products. In situ and laboratory testing was undertaken to assess the geotechnical properties of the resultant composite materials. The test pits were subsequently excavated to assess resin permeation. Drilling and resin injection was easiest through clean limestone fill materials. Recycled building waste fill material proved difficult to inject with resin; this material is thus considered unsuitable for use in resin composites. Incomplete resin permeation in several of the test pits created irregular ‘blocks’ of composite. Injected resin foams significantly improve the stiffness and resistance (strength) of the un-compacted fill material. The stiffness of the treated fill material appears to be a function of the stone particle size, its associated compaction characteristics (under loose tipping) and the proportion of resin foam matrix. The type of fill material is more critical than the type of resin to the geotechnical properties of the composite materials. Resin composites can effectively support typical design imposed loads. Compared to other traditional treatment options, such as cement grouting, the use of resin composites is potentially less disruptive, particularly for sites with limited access, and thus likely to achieve significant reinstatement cost savings. The use of resin composites is considered a suitable option for the future treatment of shallow mining collapses.Keywords: composite material, ground improvement, mining legacy, resin
Procedia PDF Downloads 35525579 Data Quality as a Pillar of Data-Driven Organizations: Exploring the Benefits of Data Mesh
Authors: Marc Bachelet, Abhijit Kumar Chatterjee, José Manuel Avila
Abstract:
Data quality is a key component of any data-driven organization. Without data quality, organizations cannot effectively make data-driven decisions, which often leads to poor business performance. Therefore, it is important for an organization to ensure that the data they use is of high quality. This is where the concept of data mesh comes in. Data mesh is an organizational and architectural decentralized approach to data management that can help organizations improve the quality of data. The concept of data mesh was first introduced in 2020. Its purpose is to decentralize data ownership, making it easier for domain experts to manage the data. This can help organizations improve data quality by reducing the reliance on centralized data teams and allowing domain experts to take charge of their data. This paper intends to discuss how a set of elements, including data mesh, are tools capable of increasing data quality. One of the key benefits of data mesh is improved metadata management. In a traditional data architecture, metadata management is typically centralized, which can lead to data silos and poor data quality. With data mesh, metadata is managed in a decentralized manner, ensuring accurate and up-to-date metadata, thereby improving data quality. Another benefit of data mesh is the clarification of roles and responsibilities. In a traditional data architecture, data teams are responsible for managing all aspects of data, which can lead to confusion and ambiguity in responsibilities. With data mesh, domain experts are responsible for managing their own data, which can help provide clarity in roles and responsibilities and improve data quality. Additionally, data mesh can also contribute to a new form of organization that is more agile and adaptable. By decentralizing data ownership, organizations can respond more quickly to changes in their business environment, which in turn can help improve overall performance by allowing better insights into business as an effect of better reports and visualization tools. Monitoring and analytics are also important aspects of data quality. With data mesh, monitoring, and analytics are decentralized, allowing domain experts to monitor and analyze their own data. This will help in identifying and addressing data quality problems in quick time, leading to improved data quality. Data culture is another major aspect of data quality. With data mesh, domain experts are encouraged to take ownership of their data, which can help create a data-driven culture within the organization. This can lead to improved data quality and better business outcomes. Finally, the paper explores the contribution of AI in the coming years. AI can help enhance data quality by automating many data-related tasks, like data cleaning and data validation. By integrating AI into data mesh, organizations can further enhance the quality of their data. The concepts mentioned above are illustrated by AEKIDEN experience feedback. AEKIDEN is an international data-driven consultancy that has successfully implemented a data mesh approach. By sharing their experience, AEKIDEN can help other organizations understand the benefits and challenges of implementing data mesh and improving data quality.Keywords: data culture, data-driven organization, data mesh, data quality for business success
Procedia PDF Downloads 13525578 Delineation of Different Geological Interfaces Beneath the Bengal Basin: Spectrum Analysis and 2D Density Modeling of Gravity Data
Authors: Md. Afroz Ansari
Abstract:
The Bengal basin is a spectacular example of a peripheral foreland basin formed by the convergence of the Indian plate beneath the Eurasian and Burmese plates. The basin is embraced on three sides; north, west and east by different fault-controlled tectonic features whereas released in the south where the rivers are drained into the Bay of Bengal. The Bengal basin in the eastern part of the Indian subcontinent constitutes the largest fluvio-deltaic to shallow marine sedimentary basin in the world today. This continental basin coupled with the offshore Bengal Fan under the Bay of Bengal forms the biggest sediment dispersal system. The continental basin is continuously receiving the sediments by the two major rivers Ganga and Brahmaputra (known as Jamuna in Bengal), and Meghna (emerging from the point of conflux of the Ganga and Brahmaputra) and large number of rain-fed, small tributaries originating from the eastern Indian Shield. The drained sediments are ultimately delivered into the Bengal fan. The significance of the present study is to delineate the variations in thicknesses of the sediments, different crustal structures, and the mantle lithosphere throughout the onshore-offshore Bengal basin. In the present study, the different crustal/geological units and the shallower mantle lithosphere were delineated by analyzing the Bouguer Gravity Anomaly (BGA) data along two long traverses South-North (running from Bengal fan cutting across the transition offshore-onshore of the Bengal basin and intersecting the Main Frontal Thrust of India-Himalaya collision zone in Sikkim-Bhutan Himalaya) and West-East (running from the Peninsular Indian Shield across the Bengal basin to the Chittagong–Tripura Fold Belt). The BGA map was derived from the analysis of topex data after incorporating Bouguer correction and all terrain corrections. The anomaly map was compared with the available ground gravity data in the western Bengal basin and the sub-continents of India for consistency of the data used. Initially, the anisotropy associated with the thicknesses of the different crustal units, crustal interfaces and moho boundary was estimated through spectral analysis of the gravity data with varying window size over the study area. The 2D density sections along the traverses were finalized after a number of iterations with the acceptable root mean square (RMS) errors. The estimated thicknesses of the different crustal units and dips of the Moho boundary along both the profiles are consistent with the earlier results. Further the results were encouraged by examining the earthquake database and focal mechanism solutions for better understanding the geodynamics. The earthquake data were taken from the catalogue of US Geological Survey, and the focal mechanism solutions were compiled from the Harvard Centroid Moment Tensor Catalogue. The concentrations of seismic events at different depth levels are not uncommon. The occurrences of earthquakes may be due to stress accumulation as a result of resistance from three sides.Keywords: anisotropy, interfaces, seismicity, spectrum analysis
Procedia PDF Downloads 27225577 Big Data Analysis with RHadoop
Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim
Abstract:
It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop
Procedia PDF Downloads 43725576 A Mutually Exclusive Task Generation Method Based on Data Augmentation
Authors: Haojie Wang, Xun Li, Rui Yin
Abstract:
In order to solve the memorization overfitting in the meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels, so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to exponential growth of computation, this paper also proposes a key data extraction method, that only extracts part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.Keywords: data augmentation, mutex task generation, meta-learning, text classification.
Procedia PDF Downloads 9325575 Ending Wars Over Water: Evaluating the Extent to Which Artificial Intelligence Can Be Used to Predict and Prevent Transboundary Water Conflicts
Authors: Akhila Potluru
Abstract:
Worldwide, more than 250 bodies of water are transboundary, meaning they cross the political boundaries of multiple countries. This creates a system of hydrological, economic, and social interdependence between communities reliant on these water sources. Transboundary water conflicts can occur as a result of this intense interdependence. Many factors contribute to the sparking of transboundary water conflicts, ranging from natural hydrological factors to hydro-political interactions. Previous attempts to predict transboundary water conflicts by analysing changes or trends in the contributing factors have typically failed because patterns in the data are hard to identify. However, there is potential for artificial intelligence and machine learning to fill this gap and identify future ‘hotspots’ up to a year in advance by identifying patterns in data where humans can’t. This research determines the extent to which AI can be used to predict and prevent transboundary water conflicts. This is done via a critical literature review of previous case studies and datasets where AI was deployed to predict water conflict. This research not only delivered a more nuanced understanding of previously undervalued factors that contribute toward transboundary water conflicts (in particular, culture and disinformation) but also by detecting conflict early, governance bodies can engage in processes to de-escalate conflict by providing pre-emptive solutions. Looking forward, this gives rise to significant policy implications and water-sharing agreements, which may be able to prevent water conflicts from developing into wide-scale disasters. Additionally, AI can be used to gain a fuller picture of water-based conflicts in areas where security concerns mean it is not possible to have staff on the ground. Therefore, AI enhances not only the depth of our knowledge about transboundary water conflicts but also the breadth of our knowledge. With demand for water constantly growing, competition between countries over shared water will increasingly lead to water conflict. There has never been a more significant time for us to be able to accurately predict and take precautions to prevent global water conflicts.Keywords: artificial intelligence, machine learning, transboundary water conflict, water management
Procedia PDF Downloads 10525574 Efficient Positioning of Data Aggregation Point for Wireless Sensor Network
Authors: Sifat Rahman Ahona, Rifat Tasnim, Naima Hassan
Abstract:
Data aggregation is a helpful technique for reducing the data communication overhead in wireless sensor network. One of the important tasks of data aggregation is positioning of the aggregator points. There are a lot of works done on data aggregation. But, efficient positioning of the aggregators points is not focused so much. In this paper, authors are focusing on the positioning or the placement of the aggregation points in wireless sensor network. Authors proposed an algorithm to select the aggregators positions for a scenario where aggregator nodes are more powerful than sensor nodes.Keywords: aggregation point, data communication, data aggregation, wireless sensor network
Procedia PDF Downloads 15725573 Bank Filtration System in Highly Mineralized Groundwater
Authors: Medalson Ronghang, Pranjal Barman, Heemantajeet Medhi
Abstract:
Bank filtration (BF) being a natural method of abstracting surface water from the river or lake via sub-surface. It can be intensively used and operated under various operating conditions for sustainability. Field investigations were carried out at various location of Kokrajhar (Assam) and Srinagar (Uttarakhand) to assess the ground water and their bank filtration wells to compare and characterized the quality. Results obtained from the analysis of the data suggest that major water quality parameter were much below the drinking water standard of BIS 10500 (2012). However, the iron concentration was found to be more than permissible limit in more than 50% of the sampled hand pump; the concentration ranged between 0.33-3.50 mg/L with acidic in nature (5.4 to 7.4) in Kokrajhar and high nitrate in Srinagar. But the abstracted water from the RBF wells has attenuated water quality with no iron concentration in Kokrajhar. The aquifers and riverbed material collected along the bank of Rivers Gaurang and Alaknanda were sieved and classified as coarse silt to medium gravel. The hydraulic conductivity was estimated in the range 5×10⁻³ to 1.4×10⁻²- 3.09×10⁻⁴-1.29 ×10⁻³ for Kokrajhar and Srinagar respectively suggesting a good permeability of the aquifer. The maximum safe yield of the well was estimated to be in the range of 4000 to 7500 L/min. This paper aims at demonstrating bank filtration method as an alternative to mineralized groundwater for drinking water.Keywords: Riverbank filtration, mineralization, water quality, groundwater
Procedia PDF Downloads 22025572 Spatial Econometric Approaches for Count Data: An Overview and New Directions
Authors: Paula Simões, Isabel Natário
Abstract:
This paper reviews a number of theoretical aspects for implementing an explicit spatial perspective in econometrics for modelling non-continuous data, in general, and count data, in particular. It provides an overview of the several spatial econometric approaches that are available to model data that are collected with reference to location in space, from the classical spatial econometrics approaches to the recent developments on spatial econometrics to model count data, in a Bayesian hierarchical setting. Considerable attention is paid to the inferential framework, necessary for structural consistent spatial econometric count models, incorporating spatial lag autocorrelation, to the corresponding estimation and testing procedures for different assumptions, to the constrains and implications embedded in the various specifications in the literature. This review combines insights from the classical spatial econometrics literature as well as from hierarchical modeling and analysis of spatial data, in order to look for new possible directions on the processing of count data, in a spatial hierarchical Bayesian econometric context.Keywords: spatial data analysis, spatial econometrics, Bayesian hierarchical models, count data
Procedia PDF Downloads 59325571 A NoSQL Based Approach for Real-Time Managing of Robotics's Data
Authors: Gueidi Afef, Gharsellaoui Hamza, Ben Ahmed Samir
Abstract:
This paper deals with the secret of the continual progression data that new data management solutions have been emerged: The NoSQL databases. They crossed several areas like personalization, profile management, big data in real-time, content management, catalog, view of customers, mobile applications, internet of things, digital communication and fraud detection. Nowadays, these database management systems are increasing. These systems store data very well and with the trend of big data, a new challenge’s store demands new structures and methods for managing enterprise data. The new intelligent machine in the e-learning sector, thrives on more data, so smart machines can learn more and faster. The robotics are our use case to focus on our test. The implementation of NoSQL for Robotics wrestle all the data they acquire into usable form because with the ordinary type of robotics; we are facing very big limits to manage and find the exact information in real-time. Our original proposed approach was demonstrated by experimental studies and running example used as a use case.Keywords: NoSQL databases, database management systems, robotics, big data
Procedia PDF Downloads 35425570 Geological Engineering Mapping Approach to Know Factor of Safety Distribution and Its Implication to Landslide Potential at Muria Mountain, Kudus, Central Java Province, Indonesia
Authors: Sony Hartono, Azka Decana, Vilia Yohana, Annisa Luthfianihuda, Yuni Faizah, Tati Andriani, Dewi Kania, Fachri Zulfiqar, Sugiar Yusup, Arman Nugraha
Abstract:
Landslide is a geological hazard that is quite common in some areas in Indonesia and have disadvantages impact for public around. Due to the high frequency of landslides in Indonesia, and extensive damage, landslides should be specifically noted. Landslides caused by a soil or rock unit that has been in a state of unstable slopes and not in ideal state again, so the value of ground resistance or the rock been passed by the value of the forces acting on the slope. Based on this fact, authors held a geological engineering mapping at Muria Mountain, Kudus, Central Java province which is known as an agriculture and religion tourism area. This geological engineering mapping is performed to determine landslides potential at Muria Mountain. Slopes stability will be illustrated by a number called the “factor of safety” where the number can describe how much potential a slope to fall. Slopes stability can be different depending on the physical and mechanical characteristics of the soil and slope conditions. Testing of physical and mechanical characteristics of the soil conducted in the geotechnical laboratory. The characteristics of the soil must be same when sampled as well as in the test laboratory. To meet that requirement, authors used "undisturb sample" method that will be guarantee sample will not be distracted by environtment influences. From laboratory tests on soil physical and mechanical properties obtained characteristics of the soil on a slope, and then inserted into a Geological Information Software that would generate a value of factor of safety and give a visualization slope form area of research. Then, as a result of the study, obtained a map of the ground movement distribution map and i is implications for landslides potential areas.Keywords: factor of safety, geological engineering mapping, landslides, slope stability, soil
Procedia PDF Downloads 41925569 Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis
Authors: C. B. Le, V. N. Pham
Abstract:
In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods.Keywords: clustering ensemble, multi-source, multi-objective, fuzzy clustering
Procedia PDF Downloads 18925568 The Revitalization of South-south Cooperation: Evaluation of South African Direct Investment in Cameroon
Authors: Albert Herve Nkolo Mpoko
Abstract:
The Foreign Direct Investment (FDI) landscape in Cameroon has garnered significant attention from both European and Asian nations due to perceived benefits such as capital infusion, technology transfer, and potential for economic expansion. However, it is noteworthy that South Africa's investment presence remains comparatively subdued in Cameroon, lagging behind that of Europe and Asia. Equally surprising is the limited footprint of Africa's economic powerhouse within other African economies. This study delved into four specific facets of South African investment in Cameroon. Initially, it focused on identifying South African companies operating within Cameroon. Subsequently, the analysis encompassed assessing the correlation between South African investment and poverty alleviation. Additionally, the study examined the nexus between South African investment and technological advancement, and underscored the significance of investment incentives in both countries Key findings of the research shed light on several crucial points. South Africa ought to reassess its economic engagement with Francophone Africa, particularly Cameroon. Despite existing policies aimed at fostering investment, there remains substantial ground to cover in this realm. The proliferation of South African enterprises in Cameroon holds the potential to ameliorate poverty and foster employment opportunities across both nations. The advent of South African firms in Cameroon can catalyse technological advancements within the region. Data collection involved surveying 100 executives from the respective administrations and conducting ten interviews. The gathered data underwent triangulation, wherein quantitative findings were juxtaposed with qualitative insights. In conclusion, the study underscores the underutilization of Cameroon by South Africa, emphasizing the untapped potential for mutual economic growth. Furthermore, it posits that the success of South Africa's multinational corporations abroad could serve as a pivotal pillar for sustaining its domestic economy.Keywords: FDI, transfer of technology, South-South cooperation, mutual economic growth
Procedia PDF Downloads 4625567 Modeling Activity Pattern Using XGBoost for Mining Smart Card Data
Authors: Eui-Jin Kim, Hasik Lee, Su-Jin Park, Dong-Kyu Kim
Abstract:
Smart-card data are expected to provide information on activity pattern as an alternative to conventional person trip surveys. The focus of this study is to propose a method for training the person trip surveys to supplement the smart-card data that does not contain the purpose of each trip. We selected only available features from smart card data such as spatiotemporal information on the trip and geographic information system (GIS) data near the stations to train the survey data. XGboost, which is state-of-the-art tree-based ensemble classifier, was used to train data from multiple sources. This classifier uses a more regularized model formalization to control the over-fitting and show very fast execution time with well-performance. The validation results showed that proposed method efficiently estimated the trip purpose. GIS data of station and duration of stay at the destination were significant features in modeling trip purpose.Keywords: activity pattern, data fusion, smart-card, XGboost
Procedia PDF Downloads 24625566 A Mutually Exclusive Task Generation Method Based on Data Augmentation
Authors: Haojie Wang, Xun Li, Rui Yin
Abstract:
In order to solve the memorization overfitting in the model-agnostic meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to an exponential growth of computation, this paper also proposes a key data extraction method that only extract part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.Keywords: mutex task generation, data augmentation, meta-learning, text classification.
Procedia PDF Downloads 14325565 Simultaneous Measurement of Wave Pressure and Wind Speed with the Specific Instrument and the Unit of Measurement Description
Authors: Branimir Jurun, Elza Jurun
Abstract:
The focus of this paper is the description of an instrument called 'Quattuor 45' and defining of wave pressure measurement. Special attention is given to measurement of wave pressure created by the wind speed increasing obtained with the instrument 'Quattuor 45' in the investigated area. The study begins with respect to theoretical attitudes and numerous up to date investigations related to the waves approaching the coast. The detailed schematic view of the instrument is enriched with pictures from ground plan and side view. Horizontal stability of the instrument is achieved by mooring which relies on two concrete blocks. Vertical wave peak monitoring is ensured by one float above the instrument. The synthesis of horizontal stability and vertical wave peak monitoring allows to create a representative database for wave pressure measuring. Instrument ‘Quattuor 45' is named according to the way the database is received. Namely, the electronic part of the instrument consists of the main chip ‘Arduino', its memory, four load cells with the appropriate modules and the wind speed sensor 'Anemometers'. The 'Arduino' chip is programmed to store two data from each load cell and two data from the anemometer on SD card each second. The next part of the research is dedicated to data processing. All measured results are stored automatically in the database and after that detailed processing is carried out in the MS Excel. The result of the wave pressure measurement is synthesized by the unit of measurement kN/m². This paper also suggests a graphical presentation of the results by multi-line graph. The wave pressure is presented on the left vertical axis, while the wind speed is shown on the right vertical axis. The time of measurement is displayed on the horizontal axis. The paper proposes an algorithm for wind speed measurements showing the results for two characteristic winds in the Adriatic Sea, called 'Bura' and 'Jugo'. The first of them is the northern wind that reaches high speeds, causing low and extremely steep waves, where the pressure of the wave is relatively weak. On the other hand, the southern wind 'Jugo' has a lower speed than the northern wind, but due to its constant duration and constant speed maintenance, it causes extremely long and high waves that cause extremely high wave pressure.Keywords: instrument, measuring unit, waves pressure metering, wind seed measurement
Procedia PDF Downloads 19725564 Hybrid CNN-SAR and Lee Filtering for Enhanced InSAR Phase Unwrapping and Coherence Optimization
Authors: Hadj Sahraoui Omar, Kebir Lahcen Wahib, Bennia Ahmed
Abstract:
Interferometric Synthetic Aperture Radar (InSAR) coherence is a crucial parameter for accurately monitoring ground deformation and environmental changes. However, coherence can be degraded by various factors such as temporal decorrelation, atmospheric disturbances, and geometric misalignments, limiting the reliability of InSAR measurements (Omar Hadj‐Sahraoui and al. 2019). To address this challenge, we propose an innovative hybrid approach that combines artificial intelligence (AI) with advanced filtering techniques to optimize interferometric coherence in InSAR data. Specifically, we introduce a Convolutional Neural Network (CNN) integrated with the Lee filter to enhance the performance of radar interferometry. This hybrid method leverages the strength of CNNs to automatically identify and mitigate the primary sources of decorrelation, while the Lee filter effectively reduces speckle noise, improving the overall quality of interferograms. We develop a deep learning-based model trained on multi-temporal and multi-frequency SAR datasets, enabling it to predict coherence patterns and enhance low-coherence regions. This hybrid CNN-SAR with Lee filtering significantly reduces noise and phase unwrapping errors, leading to more precise deformation maps. Experimental results demonstrate that our approach improves coherence by up to 30% compared to traditional filtering techniques, making it a robust solution for challenging scenarios such as urban environments, vegetated areas, and rapidly changing landscapes. Our method has potential applications in geohazard monitoring, urban planning, and environmental studies, offering a new avenue for enhancing InSAR data reliability through AI-powered optimization combined with robust filtering techniques.Keywords: CNN-SAR, Lee Filter, hybrid optimization, coherence, InSAR phase unwrapping, speckle noise reduction
Procedia PDF Downloads 1025563 Revolutionizing Traditional Farming Using Big Data/Cloud Computing: A Review on Vertical Farming
Authors: Milind Chaudhari, Suhail Balasinor
Abstract:
Due to massive deforestation and an ever-increasing population, the organic content of the soil is depleting at a much faster rate. Due to this, there is a big chance that the entire food production in the world will drop by 40% in the next two decades. Vertical farming can help in aiding food production by leveraging big data and cloud computing to ensure plants are grown naturally by providing the optimum nutrients sunlight by analyzing millions of data points. This paper outlines the most important parameters in vertical farming and how a combination of big data and AI helps in calculating and analyzing these millions of data points. Finally, the paper outlines how different organizations are controlling the indoor environment by leveraging big data in enhancing food quantity and quality.Keywords: big data, IoT, vertical farming, indoor farming
Procedia PDF Downloads 17525562 Data Challenges Facing Implementation of Road Safety Management Systems in Egypt
Authors: A. Anis, W. Bekheet, A. El Hakim
Abstract:
Implementing a Road Safety Management System (SMS) in a crowded developing country such as Egypt is a necessity. Beginning a sustainable SMS requires a comprehensive reliable data system for all information pertinent to road crashes. In this paper, a survey for the available data in Egypt and validating it for using in an SMS in Egypt. The research provides some missing data, and refer to the unavailable data in Egypt, looking forward to the contribution of the scientific society, the authorities, and the public in solving the problem of missing or unreliable crash data. The required data for implementing an SMS in Egypt are divided into three categories; the first is available data such as fatality and injury rates and it is proven in this research that it may be inconsistent and unreliable, the second category of data is not available, but it may be estimated, an example of estimating vehicle cost is available in this research, the third is not available and can be measured case by case such as the functional and geometric properties of a facility. Some inquiries are provided in this research for the scientific society, such as how to improve the links among stakeholders of road safety in order to obtain a consistent, non-biased, and reliable data system.Keywords: road safety management system, road crash, road fatality, road injury
Procedia PDF Downloads 14725561 Oily Sludge Bioremediation Pilot Plant Project, Nigeria
Authors: Ime R. Udotong, Justina I. R. Udotong, Ofonime U. M. John
Abstract:
Brass terminal, one of the several crude oil and petroleum products storage/handling facilities in the Niger Delta was built in the 1980s. Activities at this site, over the years, released crude oil into this 3 m-deep, 1500 m-long canal lying adjacent to the terminal with oil floating on it and its sediment heavily polluted. To ensure effective clean-up, three major activities were planned: Site characterization, bioremediation pilot plant construction and testing and full-scale bioremediation of contaminated sediment/bank soil by land farming. The canal was delineated into 12 lots and each characterized, with reference to the floating oily phase, contaminated sediment and canal bank soil. As a result of site characterization, a pilot plant for on-site bioremediation was designed and a treatment basin constructed for carrying out pilot bioremediation test. Following a designed sampling protocol, samples from this pilot plant were collected for analysis at two laboratories as a quality assurance/quality control check. Results showed that Brass Canal upstream is contaminated with dark, thick and viscous oily film with characteristic hydrocarbon smell while downstream, thin oily film interspersed with water were observed. Sediments were observed to be dark with mixture of brownish sandy soil with TPH ranging from 17,800 mg/kg in Lot 1 to 88,500 mg/kg in Lot 12 samples. Brass Canal bank soil was observed to be sandy from ground surface to 3m, below ground surface (bgs) it was silty-sandy and brownish while subsurface soil (4-10m bgs) was sandy-clayey and whitish/grayish with typical hydrocarbon smell. Preliminary results obtained so far have been very promising but were proprietary. This project is considered, to the best of technical literature knowledge, the first large-scale on-site bioremediation project in the Niger Delta region, Nigeria.Keywords: bioremediation, contaminated sediment, land farming, oily sludge, oil terminal
Procedia PDF Downloads 45325560 Using Photogrammetric Techniques to Map the Mars Surface
Authors: Ahmed Elaksher, Islam Omar
Abstract:
For many years, Mars surface has been a mystery for scientists. Lately with the help of geospatial data and photogrammetric procedures researchers were able to capture some insights about this planet. Two of the most imperative data sources to explore Mars are the The High Resolution Imaging Science Experiment (HiRISE) and the Mars Orbiter Laser Altimeter (MOLA). HiRISE is one of six science instruments carried by the Mars Reconnaissance Orbiter, launched August 12, 2005, and managed by NASA. The MOLA sensor is a laser altimeter carried by the Mars Global Surveyor (MGS) and launched on November 7, 1996. In this project, we used MOLA-based DEMs to orthorectify HiRISE optical images for generating a more accurate and trustful surface of Mars. The MOLA data was interpolated using the kriging interpolation technique. Corresponding tie points were digitized from both datasets. These points were employed in co-registering both datasets using GIS analysis tools. In this project, we employed three different 3D to 2D transformation models. These are the parallel projection (3D affine) transformation model; the extended parallel projection transformation model; the Direct Linear Transformation (DLT) model. A set of tie-points was digitized from both datasets. These points were split into two sets: Ground Control Points (GCPs), used to evaluate the transformation parameters using least squares adjustment techniques, and check points (ChkPs) to evaluate the computed transformation parameters. Results were evaluated using the RMSEs between the precise horizontal coordinates of the digitized check points and those estimated through the transformation models using the computed transformation parameters. For each set of GCPs, three different configurations of GCPs and check points were tested, and average RMSEs are reported. It was found that for the 2D transformation models, average RMSEs were in the range of five meters. Increasing the number of GCPs from six to ten points improve the accuracy of the results with about two and half meters. Further increasing the number of GCPs didn’t improve the results significantly. Using the 3D to 2D transformation parameters provided three to two meters accuracy. Best results were reported using the DLT transformation model. However, increasing the number of GCPS didn’t have substantial effect. The results support the use of the DLT model as it provides the required accuracy for ASPRS large scale mapping standards. However, well distributed sets of GCPs is a key to provide such accuracy. The model is simple to apply and doesn’t need substantial computations.Keywords: mars, photogrammetry, MOLA, HiRISE
Procedia PDF Downloads 5725559 Big Data-Driven Smart Policing: Big Data-Based Patrol Car Dispatching in Abu Dhabi, UAE
Authors: Oualid Walid Ben Ali
Abstract:
Big Data has become one of the buzzwords today. The recent explosion of digital data has led the organization, either private or public, to a new era towards a more efficient decision making. At some point, business decided to use that concept in order to learn what make their clients tick with phrases like ‘sales funnel’ analysis, ‘actionable insights’, and ‘positive business impact’. So, it stands to reason that Big Data was viewed through green (read: money) colored lenses. Somewhere along the line, however someone realized that collecting and processing data doesn’t have to be for business purpose only, but also could be used for other purposes to assist law enforcement or to improve policing or in road safety. This paper presents briefly, how Big Data have been used in the fields of policing order to improve the decision making process in the daily operation of the police. As example, we present a big-data driven system which is sued to accurately dispatch the patrol cars in a geographic environment. The system is also used to allocate, in real-time, the nearest patrol car to the location of an incident. This system has been implemented and applied in the Emirate of Abu Dhabi in the UAE.Keywords: big data, big data analytics, patrol car allocation, dispatching, GIS, intelligent, Abu Dhabi, police, UAE
Procedia PDF Downloads 49025558 GIS-Driven Analysis for Locating Suitable Areas for Renewable Energy
Authors: Saleh Nabiyev
Abstract:
Renewable energy is becoming increasingly important in today's world due to its significant impact on the green economy, ecology, environment, and climate change. Renewable energy sources, such as solar and wind, are clean and sustainable, making them an ideal solution to reduce carbon emissions and mitigate the effects of climate change. The Karabakh region is located in the South Caucasus and covers an area of approximately 11,500 km². The region has a mountainous terrain, which can affect the availability of wind and solar resources. The Karabakh region has significant wind power potential, particularly in its mountainous areas where wind speeds are typically higher. According to a study conducted by the European Commission Joint Research Centre, the average wind speed in the Karabakh region is between 4 and 6 meters per second (m/s) at a height of 50 meters above ground level (AGL). However, wind speeds can be higher in some areas, reaching up to 10 m/s in some mountainous areas. The region also has significant solar power potential, with an average of 2,000 to 2,200 hours of sunshine per year. The region's high altitude and clear skies make it particularly suitable for the development of solar power projects. In this research, the application of satellite images, solar radiation, wind speed and direction, as well as various other materials to determine suitable areas for alternative energy sources, is investigated. The methodology for selecting suitable locations for solar and wind energy consists of four main parts: identification of factors, evaluation of factors, data preparation, and application of suitability analysis. At the end of the research, the territory of the Kalbajar and Lachin districts is suitable for wind energy. The southern plain part of Karabakh is highly evaluated in terms of solar energy potential, especially Jabrayil district. Generally, outcomes taken from this research are essential data for increasing of rational using natural resources, as well as combating climate change.Keywords: GIS, remote sensing, suitability analysis, solar energy, wind energy
Procedia PDF Downloads 3125557 Study of Radiological and Chemical Effects of Uranium in Ground Water of SW and NE Punjab, India
Authors: Komal Saini, S. K. Sahoo, B. S. Bajwa
Abstract:
The Laser Fluorimetery Technique has been used for the microanalysis of uranium content in water samples collected from different sources like the hand pumps, tube wells in the drinking water samples of SW & NE Punjab, India. The geographic location of the study region in NE Punjab is between latitude 31.21º- 32.05º N and longitude 75.60º-76.14º E and for SW Punjab is between latitude 29.66º-30.48º N and longitude 74.69º-75.54º E. The purpose of this study was mainly to investigate the uranium concentration levels of ground water being used for drinking purposes and to determine its health effects, if any, to the local population of these regions. In the present study 131 samples of drinking water collected from different villages of SW and 95 samples from NE, Punjab state, India have been analyzed for chemical and radiological toxicity. In the present investigation, uranium content in water samples of SW Punjab ranges from 0.13 to 908 μgL−1 with an average of 82.1 μgL−1 whereas in samples collected from NE- Punjab, it ranges from 0 to 28.2 μgL−1 with an average of 4.84 μgL−1. Thus, revealing that in the SW- Punjab 54 % of drinking water samples have uranium concentration higher than international recommended limit of 30 µgl-1 (WHO, 2011) while 35 % of samples exceeds the threshold of 60 µgl-1 recommended by our national regulatory authority of Atomic Energy Regulatory Board (AERB), Department of Atomic Energy, India, 2004. On the other hand in the NE-Punjab region, none of the observed water sample has uranium content above the national/international recommendations. The observed radiological risk in terms of excess cancer risk ranges from 3.64x10-7 to 2.54x10-3 for SW-Punjab, whereas for NE region it ranges from 0 to 7.89x10-5. The chemical toxic effect in terms of Life-time average Daily Dose (LDD) and Hazard Quotient (HQ) have also been calculated. The LDD for SW-Punjab varies from 0.0098 to 68.46 with an average of 6.18 µg/ kg/day whereas for NE region it varies from 0 to 2.13 with average 0.365 µg/ kg/day, thus indicating presence of chemical toxicity in SW Punjab as 35% of the observed samples in the SW Punjab are above the recommendation limit of 4.53 µg/ kg/day given by AERB for 60 µgl-1 of uranium. Maximum & Minimum values for hazard quotient for SW Punjab is 0.002 & 15.11 with average 1.36 which is considerably high as compared to safe limit i.e. 1. But for NE Punjab HQ varies from 0 to 0.47. The possible sources of high uranium observed in the SW- Punjab will also be discussed.Keywords: uranium, groundwater, radiological and chemical toxicity, Punjab, India
Procedia PDF Downloads 38125556 A Comparative Study of Regional Climate Models and Global Coupled Models over Uttarakhand
Authors: Sudip Kumar Kundu, Charu Singh
Abstract:
As a great physiographic divide, the Himalayas affecting a large system of water and air circulation which helps to determine the climatic condition in the Indian subcontinent to the south and mid-Asian highlands to the north. It creates obstacles by defending chill continental air from north side into India in winter and also defends rain-bearing southwesterly monsoon to give up maximum precipitation in that area in monsoon season. Nowadays extreme weather conditions such as heavy precipitation, cloudburst, flash flood, landslide and extreme avalanches are the regular happening incidents in the region of North Western Himalayan (NWH). The present study has been planned to investigate the suitable model(s) to find out the rainfall pattern over that region. For this investigation, selected models from Coordinated Regional Climate Downscaling Experiment (CORDEX) and Coupled Model Intercomparison Project Phase 5 (CMIP5) has been utilized in a consistent framework for the period of 1976 to 2000 (historical). The ability of these driving models from CORDEX domain and CMIP5 has been examined according to their capability of the spatial distribution as well as time series plot of rainfall over NWH in the rainy season and compared with the ground-based Indian Meteorological Department (IMD) gridded rainfall data set. It is noted from the analysis that the models like MIROC5 and MPI-ESM-LR from the both CORDEX and CMIP5 provide the best spatial distribution of rainfall over NWH region. But the driving models from CORDEX underestimates the daily rainfall amount as compared to CMIP5 driving models as it is unable to capture daily rainfall data properly when it has been plotted for time series (TS) individually for the state of Uttarakhand (UK) and Himachal Pradesh (HP). So finally it can be said that the driving models from CMIP5 are better than CORDEX domain models to investigate the rainfall pattern over NWH region.Keywords: global warming, rainfall, CMIP5, CORDEX, NWH
Procedia PDF Downloads 16925555 Mining Multicity Urban Data for Sustainable Population Relocation
Authors: Xu Du, Aparna S. Varde
Abstract:
In this research, we propose to conduct diagnostic and predictive analysis about the key factors and consequences of urban population relocation. To achieve this goal, urban simulation models extract the urban development trends as land use change patterns from a variety of data sources. The results are treated as part of urban big data with other information such as population change and economic conditions. Multiple data mining methods are deployed on this data to analyze nonlinear relationships between parameters. The result determines the driving force of population relocation with respect to urban sprawl and urban sustainability and their related parameters. Experiments so far reveal that data mining methods discover useful knowledge from the multicity urban data. This work sets the stage for developing a comprehensive urban simulation model for catering to specific questions by targeted users. It contributes towards achieving sustainability as a whole.Keywords: data mining, environmental modeling, sustainability, urban planning
Procedia PDF Downloads 30825554 Major Causes of Delay in Construction Projects
Authors: Y. Gholipour, E. Rezazadeh
Abstract:
Delay is one of the most serious and common problems of construction project that can affect project delivery unfavorably. This research presents the most important causes of delay in large dam projects based on a survey on some executed dam construction in Iran. In this survey a randomly selected samples of owners, consultants and contractors have been involved. The outcome of this survey revealed that scheduled payments, site management, shop drawing review process, unforeseen ground conditions and contractor experience as the most important factors affecting on delay in dam construction projects.Keywords: delay, dam construction, project management, Iran
Procedia PDF Downloads 444