Search results for: complexity ontology
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1789

Search results for: complexity ontology

829 Real-Time Multi-Vehicle Tracking Application at Intersections Based on Feature Selection in Combination with Color Attribution

Authors: Qiang Zhang, Xiaojian Hu

Abstract:

In multi-vehicle tracking, based on feature selection, the tracking system efficiently tracks vehicles in a video with minimal error in combination with color attribution, which focuses on presenting a simple and fast, yet accurate and robust solution to the problem such as inaccurately and untimely responses of statistics-based adaptive traffic control system in the intersection scenario. In this study, a real-time tracking system is proposed for multi-vehicle tracking in the intersection scene. Considering the complexity and application feasibility of the algorithm, in the object detection step, the detection result provided by virtual loops were post-processed and then used as the input for the tracker. For the tracker, lightweight methods were designed to extract and select features and incorporate them into the adaptive color tracking (ACT) framework. And the approbatory online feature selection algorithms are integrated on the mature ACT system with good compatibility. The proposed feature selection methods and multi-vehicle tracking method are evaluated on KITTI datasets and show efficient vehicle tracking performance when compared to the other state-of-the-art approaches in the same category. And the system performs excellently on the video sequences recorded at the intersection. Furthermore, the presented vehicle tracking system is suitable for surveillance applications.

Keywords: real-time, multi-vehicle tracking, feature selection, color attribution

Procedia PDF Downloads 153
828 Proposed Framework based on Classification of Vertical Handover Decision Strategies in Heterogeneous Wireless Networks

Authors: Shidrokh Goudarzi, Wan Haslina Hassan

Abstract:

Heterogeneous wireless networks are converging towards an all-IP network as part of the so-called next-generation network. In this paradigm, different access technologies need to be interconnected; thus, vertical handovers or vertical handoffs are necessary for seamless mobility. In this paper, we conduct a review of existing vertical handover decision-making mechanisms that aim to provide ubiquitous connectivity to mobile users. To offer a systematic comparison, we categorize these vertical handover measurement and decision structures based on their respective methodology and parameters. Subsequently, we analyze several vertical handover approaches in the literature and compare them according to their advantages and weaknesses. The paper compares the algorithms based on the network selection methods, complexity of the technologies used and efficiency in order to introduce our vertical handover decision framework. We find that vertical handovers on heterogeneous wireless networks suffer from the lack of a standard and efficient method to satisfy both user and network quality of service requirements at different levels including architectural, decision-making and protocols. Also, the consolidation of network terminal, cross-layer information, multi packet casting and intelligent network selection algorithm appears to be an optimum solution for achieving seamless service continuity in order to facilitate seamless connectivity.

Keywords: heterogeneous wireless networks, vertical handovers, vertical handover metric, decision-making algorithms

Procedia PDF Downloads 382
827 Attachment Patterns in a Sample of South African Children at Risk in Middle Childhood

Authors: Renate Gericke, Carol Long

Abstract:

Despite the robust empirical support of attachment, advancement in the description and conceptualization of attachment has been slow and has not significantly advanced beyond the identification of attachment security or type (namely, secure, avoidant, ambivalent and disorganized). This has continued despite papers arguing for theoretical refinement in the classification of attachment presentations. For thinking and practice to advance, it is critically important that these categories and their assessment be interrogated in different contexts and across developmental age. To achieve this, a quantitative design was used with descriptive and inferential statistics, and general linear models were employed to analyze the data. The Attachment Story Completion Test (ASCT) was administered to 105 children between the ages of eight and twelve from socio-economically deprived contexts with high exposure to trauma. A staggering 93% of the children had insecure attachments (specifically, avoidant 37%, disorganized 34% and ambivalent 22%) and attachment was more complex than currently conceptualized in the attachment literature. Primary attachment did not only present as one of four discreet categories, but 70% of the sample had a complex attachment with more than one type of maternal attachment style. Attachment intensity also varied along a continuum (between 1 and 5). The findings have implications for a) research that has not considered the potential complexity of attachment or attachment intensity, b) policy to more actively support mother-infant dyads, particularly in high-risk contexts and c) question the applicability of a western conceptualization of a primary maternal attachment figure in non-western collectivist societies.

Keywords: attachment, children at risk, middle childhood, non-western context

Procedia PDF Downloads 183
826 Membrane Bioreactor versus Activated Sludge Process for Aerobic Wastewater Treatment and Recycling

Authors: Sarra Kitanou

Abstract:

Membrane bioreactor (MBR) systems are one of the most widely used wastewater treatment processes for various municipal and industrial waste streams. It is based on complex interactions between biological processes, filtration process and rheological properties of the liquid to be treated. Its complexity makes understanding system operation and optimization more difficult, and traditional methods based on experimental analysis are costly and time consuming. The present study was based on an external membrane bioreactor pilot scale with ceramic membranes compared to conventional activated sludge process (ASP) plant. Both systems received their influent from a domestic wastewater. The membrane bioreactor (MBR) produced an effluent with much better quality than ASP in terms of total suspended solids (TSS), organic matter such as biological oxygen demand (BOD) and chemical oxygen demand (COD), total Phosphorus and total Nitrogen. Other effluent quality parameters also indicate substantial differences between ASP and MBR. This study leads to conclude that in the case domestic wastewater, MBR treatment has excellent effluent quality. Hence, the replacement of the ASP by the MBRs may be justified on the basis of their improved removal of solids, nutrients, and micropollutants. Furthermore, in terms of reuse the great quality of the treated water allows it to be reused for irrigation.

Keywords: aerobic wastewater treatment, conventional activated sludge process, membrane bioreactor, reuse for irrigation

Procedia PDF Downloads 67
825 Molecular Interactions Driving RNA Binding to hnRNPA1 Implicated in Neurodegeneration

Authors: Sakina Fatima, Joseph-Patrick W. E. Clarke, Patricia A. Thibault, Subha Kalyaanamoorthy, Michael Levin, Aravindhan Ganesan

Abstract:

Heteronuclear ribonucleoprotein (hnRNPA1 or A1) is associated with the pathology of different diseases, including neurological disorders and cancers. In particular, the aggregation and dysfunction of A1 have been identified as a critical driver for neurodegeneration (NDG) in Multiple Sclerosis (MS). Structurally, A1 includes a low-complexity domain (LCD) and two RNA-recognition motifs (RRMs), and their interdomain coordination may play a crucial role in A1 aggregation. Previous studies propose that RNA-inhibitors or nucleoside analogs that bind to RRMs can potentially prevent A1 self-association. Therefore, molecular-level understanding of the structures, dynamics, and nucleotide interactions with A1 RRMs can be useful for developing therapeutics for NDG in MS. In this work, a combination of computational modelling and biochemical experiments were employed to analyze a set of RNA-A1 RRM complexes. Initially, the atomistic models of RNA-RRM complexes were constructed by modifying known crystal structures (e.g., PDBs: 4YOE and 5MPG), and through molecular docking calculations. The complexes were optimized using molecular dynamics simulations (200-400 ns), and their binding free energies were computed. The binding affinities of the selected complexes were validated using a thermal shift assay. Further, the most important molecular interactions that contributed to the overall stability of the RNA-A1 RRM complexes were deduced. The results highlight that adenine and guanine are the most suitable nucleotides for high-affinity binding with A1. These insights will be useful in the rational design of nucleotide-analogs for targeting A1 RRMs.

Keywords: hnRNPA1, molecular docking, molecular dynamics, RNA-binding proteins

Procedia PDF Downloads 106
824 The Use of Water Resources Yield Model at Kleinfontein Dam

Authors: Lungile Maliba, O. I. Nkwonta, E Onyari

Abstract:

Water resources development and management are regarded as crucial for poverty reduction in many developing countries and sustainable economic growth such as South Africa. The contribution of large hydraulic infrastructure and management of it, particularly reservoirs, to development remains controversial. This controversy stems from the fact that from a historical point of view construction of reservoirs has brought fewer benefits than envisaged and has resulted in significant environmental and social costs. A further complexity in reservoir management is the variety of stakeholders involved, all with different objectives, including domestic and industrial water use, flood control, irrigation and hydropower generation. The objective was to evaluate technical adaptation options for kleinfontein Dam’s current operating rule curves. To achieve this objective, the current operating rules curves being used in the sub-basin were analysed. An objective methodology was implemented in other to get the operating rules with regards to the target storage curves. These were derived using the Water Resources Yield/Planning Model (WRY/PM), with the aim of maximising of releases to demand zones. The result showed that the system is over allocated and in addition the demands exceed the long-term yield that is available for the system. It was concluded that the current operating rules in the system do not produce the optimum operation such as target storage curves to avoid supply failures in the system.

Keywords: infrastructure, Kleinfontein dam, operating rule curve, water resources yield and planning model

Procedia PDF Downloads 131
823 The Importance of Downstream Supply Chain in Supply Chain Risk Management: Multi-Objective Optimization

Authors: Zohreh Khojasteh-Ghamari, Takashi Irohara

Abstract:

One of the efficient ways in supply chain risk management is avoiding the interruption in Supply Chain (SC) before it occurs. Although the majority of the organizations focus on their first-tier suppliers to avoid risk in the SC, studies show that in only 60 percent of the disruption cases the reason is first tier suppliers. In the 40 percent of the SC disruptions, the reason is downstream SC, which is the second tier and lower. Due to the increasing complexity and interrelation of modern supply chains, the SC elements have become difficult to trace. Moreover, studies show that there is a vital need to better understand the integration of risk and visibility, especially in the context of multiple objectives. In this study, we propose a multi-objective programming model to avoid disruption in SC. The objective of this study is evaluating the effect of downstream SCV on managing supply chain risk. We propose a multi-objective mathematical programming model with the objective functions of minimizing the total cost and maximizing the downstream supply chain visibility (SCV). The decision variable is supplier selection. We assume there are several manufacturers and several candidate suppliers. For each manufacturer, our model proposes the best suppliers with the lowest cost and maximum visibility in downstream supply chain. We examine the applicability of the model by numerical examples. We also define several scenarios for datasets and observe the tendency. The results show that minimum visibility in downstream SC is needed to have a safe SC network.

Keywords: downstream supply chain, optimization, supply chain risk, supply chain visibility

Procedia PDF Downloads 237
822 Agony and Agency: Discursive Construction of Barren women in the Bible and Traditional African Society

Authors: Vicky Khasandi-Telewa, Sinfree Makoni

Abstract:

Barrenness is a fundamentally agonizing condition that leads to identity disruption in its victims. In Africa, women are usually referred to as ‘Mother of X,’ and this causes grief to one who does not have a child to be identified with. This paper is an examination and critical appraisal of the impact of barrenness on the self-perception of women and the underlying power relations in how they are discursively constructed in the Bible and Traditional African Society (TAS). It is an analysis of expressive practices to examine how barrenness is constructed in Christianity and TAS with the aim of understanding the intersecting power systems. We approach this from an integrationism and Critical Discourse Analysis perspective that takes seriously both the radical harassment of barren women and the possibilities offered by the ensuing desperation calling for inclusive reinterpretation. We also seek to understand barren women’s coping mechanisms and suggestions on how best to improve their lives. The purpose of this study is to explain how discursive construction of barrenness affects the fundamental rights and freedoms of women and what linguistic strategies they adopt to navigate through the maze of stigma. It seeks to illustrate a more nuanced complexity of barren women's lives through women's own exegesis of the Biblical accounts of barrenness and their traditions and to explore alternative narratives. We explore the linguistic strategies the barren women employ to communicate their coping with limitations imposed upon their rights by the negative constructions.

Keywords: integrationism, critical discourse analysis, barrenness, communication strategies, women rights

Procedia PDF Downloads 56
821 D3Advert: Data-Driven Decision Making for Ad Personalization through Personality Analysis Using BiLSTM Network

Authors: Sandesh Achar

Abstract:

Personalized advertising holds greater potential for higher conversion rates compared to generic advertisements. However, its widespread application in the retail industry faces challenges due to complex implementation processes. These complexities impede the swift adoption of personalized advertisement on a large scale. Personalized advertisement, being a data-driven approach, necessitates consumer-related data, adding to its complexity. This paper introduces an innovative data-driven decision-making framework, D3Advert, which personalizes advertisements by analyzing personalities using a BiLSTM network. The framework utilizes the Myers–Briggs Type Indicator (MBTI) dataset for development. The employed BiLSTM network, specifically designed and optimized for D3Advert, classifies user personalities into one of the sixteen MBTI categories based on their social media posts. The classification accuracy is 86.42%, with precision, recall, and F1-Score values of 85.11%, 84.14%, and 83.89%, respectively. The D3Advert framework personalizes advertisements based on these personality classifications. Experimental implementation and performance analysis of D3Advert demonstrate a 40% improvement in impressions. D3Advert’s innovative and straightforward approach has the potential to transform personalized advertising and foster widespread personalized advertisement adoption in marketing.

Keywords: personalized advertisement, deep Learning, MBTI dataset, BiLSTM network, NLP.

Procedia PDF Downloads 31
820 Extended Literature Review on Sustainable Energy by Using Multi-Criteria Decision Making Techniques

Authors: Koray Altintas, Ozalp Vayvay

Abstract:

Increased global issues such as depletion of sources, environmental problems and social inequality triggered public awareness towards finding sustainable solutions in order to ensure the well-being of the current as well as future generations. Since energy plays a significant role in improved social and economic well-being and is imperative on both industrial and commercial wealth creation, it is a must to develop a standardized set of metrics which makes it possible to indicate the present condition relative to conditions in the past and to develop any perspective which is required to frame actions for the future. This is not an easy task by considering the complexity of the issue which requires integrating economic, environmental and social aspects of sustainable energy. Multi-criteria decision making (MCDM) can be considered as a form of integrated sustainability evaluation and a decision support approach that can be used to solve complex problems featuring; conflicting objectives, different forms of data and information, multi-interests and perspectives. On that matter, MCDM methods are useful for providing solutions to complex energy management problems. The aim of this study is to review MCDM approaches that can be used for examining sustainable energy management. This study presents an insight into MCDM techniques and methods that can be useful for engineers, researchers and policy makers working in the energy sector.

Keywords: sustainable energy, sustainability criteria, multi-criteria decision making, sustainability dimensions

Procedia PDF Downloads 309
819 Ground Deformation Module for the New Laboratory Methods

Authors: O. Giorgishvili

Abstract:

For calculation of foundations one of the important characteristics is the module of deformation (E0). As we all know, the main goal of calculation of the foundations of buildings on deformation is to arrange the base settling and difference in settlings in such limits that do not cause origination of cracks and changes in design levels that will be dangerous to standard operation in the buildings and their individual structures. As is known from the literature and the practical application, the modulus of deformation is determined by two basic methods: laboratory method, soil test on compression (without the side widening) and soil test in field conditions. As we know, the deformation modulus of soil determined by field method is closer to the actual modulus deformation of soil, but the complexity of the tests to be carried out and the financial concerns did not allow determination of ground deformation modulus by field method. Therefore, we determine the ground modulus of deformation by compression method without side widening. Concerning this, we introduce a new way for determination of ground modulus of deformation by laboratory order that occurs by side widening and more accurately reflects the ground modulus of deformation and more accurately reflects the actual modulus of deformation and closer to the modulus of deformation determined by the field method. In this regard, we bring a new approach on the ground deformation detection laboratory module, which is done by widening sides. The tests and the results showed that the proposed method of ground deformation modulus is closer to the results that are obtained in the field, which reflects the foundation's work in real terms more accurately than the compression of the ground deformation module.

Keywords: build, deformation modulus, foundations, ground, laboratory research

Procedia PDF Downloads 361
818 A Study on Design for Parallel Test Based on Embedded System

Authors: Zheng Sun, Weiwei Cui, Xiaodong Ma, Hongxin Jin, Dongpao Hong, Jinsong Yang, Jingyi Sun

Abstract:

With the improvement of the performance and complexity of modern equipment, automatic test system (ATS) becomes widely used for condition monitoring and fault diagnosis. However, the conventional ATS mainly works in a serial mode, and lacks the ability of testing several equipments at the same time. That leads to low test efficiency and ATS redundancy. Especially for a large majority of equipment under test, the conventional ATS cannot meet the requirement of efficient testing. To reduce the support resource and increase test efficiency, we propose a method of design for the parallel test based on the embedded system in this paper. Firstly, we put forward the general framework of the parallel test system, and the system contains a central management system (CMS) and several distributed test subsystems (DTS). Then we give a detailed design of the system. For the hardware of the system, we use embedded architecture to design DTS. For the software of the system, we use test program set to improve the test adaption. By deploying the parallel test system, the time to test five devices is now equal to the time to test one device in the past. Compared with the conventional test system, the proposed test system reduces the size and improves testing efficiency. This is of great significance for equipment to be put into operation swiftly. Finally, we take an industrial control system as an example to verify the effectiveness of the proposed method. The result shows that the method is reasonable, and the efficiency is improved up to 500%.

Keywords: parallel test, embedded system, automatic test system, automatic test system (ATS), central management system, central management system (CMS), distributed test subsystems, distributed test subsystems (DTS)

Procedia PDF Downloads 287
817 Automatic Tuning for a Systemic Model of Banking Originated Losses (SYMBOL) Tool on Multicore

Authors: Ronal Muresano, Andrea Pagano

Abstract:

Nowadays, the mathematical/statistical applications are developed with more complexity and accuracy. However, these precisions and complexities have brought as result that applications need more computational power in order to be executed faster. In this sense, the multicore environments are playing an important role to improve and to optimize the execution time of these applications. These environments allow us the inclusion of more parallelism inside the node. However, to take advantage of this parallelism is not an easy task, because we have to deal with some problems such as: cores communications, data locality, memory sizes (cache and RAM), synchronizations, data dependencies on the model, etc. These issues are becoming more important when we wish to improve the application’s performance and scalability. Hence, this paper describes an optimization method developed for Systemic Model of Banking Originated Losses (SYMBOL) tool developed by the European Commission, which is based on analyzing the application's weakness in order to exploit the advantages of the multicore. All these improvements are done in an automatic and transparent manner with the aim of improving the performance metrics of our tool. Finally, experimental evaluations show the effectiveness of our new optimized version, in which we have achieved a considerable improvement on the execution time. The time has been reduced around 96% for the best case tested, between the original serial version and the automatic parallel version.

Keywords: algorithm optimization, bank failures, OpenMP, parallel techniques, statistical tool

Procedia PDF Downloads 356
816 Glycan Analyzer: Software to Annotate Glycan Structures from Exoglycosidase Experiments

Authors: Ian Walsh, Terry Nguyen-Khuong, Christopher H. Taron, Pauline M. Rudd

Abstract:

Glycoproteins and their covalently bonded glycans play critical roles in the immune system, cell communication, disease and disease prognosis. Ultra performance liquid chromatography (UPLC) coupled with mass spectrometry is conventionally used to qualitatively and quantitatively characterise glycan structures in a given sample. Exoglycosidases are enzymes that catalyze sequential removal of monosaccharides from the non-reducing end of glycans. They naturally have specificity for a particular type of sugar, its stereochemistry (α or β anomer) and its position of attachment to an adjacent sugar on the glycan. Thus, monitoring the peak movements (both in the UPLC and MS1) after application of exoglycosidases provides a unique and effective way to annotate sugars with high detail - i.e. differentiating positional and linkage isomers. Manual annotation of an exoglycosidase experiment is difficult and time consuming. As such, with increasing sample complexity and the number of exoglycosidases, the analysis could result in manually interpreting hundreds of peak movements. Recently, we have implemented pattern recognition software for automated interpretation of UPLC-MS1 exoglycosidase digestions. In this work, we explain the software, indicate how much time it will save and provide example usage showing the annotation of positional and linkage isomers in Immunoglobulin G, apolipoprotein J, and simple glycan standards.

Keywords: bioinformatics, automated glycan assignment, liquid chromatography, mass spectrometry

Procedia PDF Downloads 185
815 Power Allocation Algorithm for Orthogonal Frequency Division Multiplexing Based Cognitive Radio Networks

Authors: Bircan Demiral

Abstract:

Cognitive radio (CR) is the promising technology that addresses the spectrum scarcity problem for future wireless communications. Orthogonal Frequency Division Multiplexing (OFDM) technology provides more power band ratios for cognitive radio networks (CRNs). While CR is a solution to the spectrum scarcity, it also brings up the capacity problem. In this paper, a novel power allocation algorithm that aims at maximizing the sum capacity in the OFDM based cognitive radio networks is proposed. Proposed allocation algorithm is based on the previously developed water-filling algorithm. To reduce the computational complexity calculating in water filling algorithm, proposed algorithm allocates the total power according to each subcarrier. The power allocated to the subcarriers increases sum capacity. To see this increase, Matlab program was used, and the proposed power allocation was compared with average power allocation, water filling and general power allocation algorithms. The water filling algorithm performed worse than the proposed algorithm while it performed better than the other two algorithms. The proposed algorithm is better than other algorithms in terms of capacity increase. In addition the effect of the change in the number of subcarriers on capacity was discussed. Simulation results show that the increase in the number of subcarrier increases the capacity.

Keywords: cognitive radio network, OFDM, power allocation, water filling

Procedia PDF Downloads 125
814 Partnership Brokering as a Driver of Social Business

Authors: Lani Fraizer, Faiz Shah

Abstract:

Extreme poverty continues to plague the world. Forty-seven million people live well-below the poverty line in Bangladesh, enduring poor quality of life, often with no access to basic human needs like shelter and healthcare. It is not surprising that poverty eradication is central to the mission of social change makers, such as Muhammad Yunus, who have demonstrated how enterprise-led development initiatives empower individuals at the grassroots, and can galvanize entire communities to emerge out of poverty. Such strategies call for system-wide change, and like a number of systems leaders, social business champions have typically challenged the status quo, and broken out of silos to catalyze vibrant multi-stakeholder partnerships across sectors. Apart from individual charisma, social change makers succeed because they garner collaborative impact through socially beneficial partnerships. So while enterprise-led social development evolves in scope and complexity, in step with the need to create and sustain partnerships, Partnership Brokering is emerging as an approach to facilitate collaborative processes. As such, it may now be possible for anyone motivated by the idea of social business to acquire the skills and sophistication necessary for building enriching partnerships that harness the power of the market to address poverty. This paper examines dimensions of partnership brokering in the context of social business, and explores the implications of this emerging approach on fostering poverty eradication.

Keywords: poverty, social business, partnership brokering, social entrepreneurship, systems change, enterprise-led development, change making

Procedia PDF Downloads 243
813 Transfer of Constraints or Constraints on Transfer? Syntactic Islands in Danish L2 English

Authors: Anne Mette Nyvad, Ken Ramshøj Christensen

Abstract:

In the syntax literature, it has standardly been assumed that relative clauses and complement wh-clauses are islands for extraction in English, and that constraints on extraction from syntactic islands are universal. However, the Mainland Scandinavian languages has been known to provide counterexamples. Previous research on Danish has shown that neither relative clauses nor embedded questions are strong islands in Danish. Instead, extraction from this type of syntactic environment is degraded due to structural complexity and it interacts with nonstructural factors such as the frequency of occurrence of the matrix verb, the possibility of temporary misanalysis leading to semantic incongruity and exposure over time. We argue that these facts can be accounted for with parametric variation in the availability of CP-recursion, resulting in the patterns observed, as Danish would then “suspend” the ban on movement out of relative clauses and embedded questions. Given that Danish does not seem to adhere to allegedly universal syntactic constraints, such as the Complex NP Constraint and the Wh-Island Constraint, what happens in L2 English? We present results from a study investigating how native Danish speakers judge extractions from island structures in L2 English. Our findings suggest that Danes transfer their native language parameter setting when asked to judge island constructions in English. This is compatible with the Full Transfer Full Access Hypothesis, as the latter predicts that Danish would have difficulties resetting their [+/- CP-recursion] parameter in English because they are not exposed to negative evidence.

Keywords: syntax, islands, second language acquisition, danish

Procedia PDF Downloads 111
812 Identifying and Ranking Environmental Risks of Oil and Gas Projects Using the VIKOR Method for Multi-Criteria Decision Making

Authors: Sasan Aryaee, Mahdi Ravanshadnia

Abstract:

Naturally, any activity is associated with risk, and humans have understood this concept from very long times ago and seek to identify its factors and sources. On the one hand, proper risk management can cause problems such as delays and unforeseen costs in the development projects, temporary or permanent loss of services, getting lost or information theft, complexity and limitations in processes, unreliable information caused by rework, holes in the systems and many such problems. In the present study, a model has been presented to rank the environmental risks of oil and gas projects. The statistical population of the study consists of all executives active in the oil and gas fields, that the statistical sample is selected randomly. In the framework of the proposed method, environmental risks of oil and gas projects were first extracted, then a questionnaire based on these indicators was designed based on Likert scale and distributed among the statistical sample. After assessing the validity and reliability of the questionnaire, environmental risks of oil and gas projects were ranked using the VIKOR method of multiple-criteria decision-making. The results showed that the best options for HSE planning of oil and gas projects that caused the reduction of risks and personal injury and casualties and less than other options is costly for the project and it will add less time to the duration of implementing the project is the entering of dye to the environment when painting the generator pond and the presence of the rigger near the crane.

Keywords: ranking, multi-criteria decision making, oil and gas projects, HSEmanagement, environmental risks

Procedia PDF Downloads 145
811 Reduce the Impact of Wildfires by Identifying Them Early from Space and Sending Location Directly to Closest First Responders

Authors: Gregory Sullivan

Abstract:

The evolution of global warming has escalated the number and complexity of forest fires around the world. As an example, the United States and Brazil combined generated more than 30,000 forest fires last year. The impact to our environment, structures and individuals is incalculable. The world has learned to try to take this in stride, trying multiple ways to contain fires. Some countries are trying to use cameras in limited areas. There are discussions of using hundreds of low earth orbit satellites and linking them together, and, interfacing them through ground networks. These are all truly noble attempts to defeat the forest fire phenomenon. But there is a better, simpler answer. A bigger piece of the solutions puzzle is to see the fires while they are small, soon after initiation. The approach is to see the fires while they are very small and report their location (latitude and longitude) to local first responders. This is done by placing a sensor at geostationary orbit (GEO: 26,000 miles above the earth). By placing this small satellite in GEO, we can “stare” at the earth, and sense temperature changes. We do not “see” fires, but “measure” temperature changes. This has already been demonstrated on an experimental scale. Fires were seen at close to initiation, and info forwarded to first responders. it were the first to identify the fires 7 out of 8 times. The goal is to have a small independent satellite at GEO orbit focused only on forest fire initiation. Thus, with one small satellite, focused only on forest fire initiation, we hope to greatly decrease the impact to persons, property and the environment.

Keywords: space detection, wildfire early warning, demonstration wildfire detection and action from space, space detection to first responders

Procedia PDF Downloads 55
810 A Multi-Scale Approach for the Analysis of Fiber-Reinforced Composites

Authors: Azeez Shaik, Amit Salvi, B. P. Gautham

Abstract:

Fiber reinforced polymer resin composite materials are finding wide variety of applications in automotive and aerospace industry because of their high specific stiffness and specific strengths when compared to metals. New class of 2D and 3D textile and woven fabric composites offer excellent fracture toughens as they bridge the cracks formed during fracture. Due to complexity of their fiber architectures and its resulting composite microstructures, optimized design and analysis of these structures is very complicated. A traditional homogenization approach is typically used to analyze structures made up of these materials. This approach usually fails to predict damage initiation as well as damage propagation and ultimate failure of structure made up of woven and textile composites. This study demonstrates a methodology to analyze woven and textile composites by using the multi-level multi-scale modelling approach. In this approach, a geometric repetitive unit cell (RUC) is developed with all its constituents to develop a representative volume element (RVE) with all its constituents and their interaction modeled correctly. The structure is modeled based on the RUC/RVE and analyzed at different length scales with desired levels of fidelity incorporating the damage and failure. The results are passed across (up and down) the scales qualitatively as well as quantitatively from the perspective of material, configuration and architecture.

Keywords: cohesive zone, multi-scale modeling, rate dependency, RUC, woven textiles

Procedia PDF Downloads 354
809 Finite-Sum Optimization: Adaptivity to Smoothness and Loopless Variance Reduction

Authors: Bastien Batardière, Joon Kwon

Abstract:

For finite-sum optimization, variance-reduced gradient methods (VR) compute at each iteration the gradient of a single function (or of a mini-batch), and yet achieve faster convergence than SGD thanks to a carefully crafted lower-variance stochastic gradient estimator that reuses past gradients. Another important line of research of the past decade in continuous optimization is the adaptive algorithms such as AdaGrad, that dynamically adjust the (possibly coordinate-wise) learning rate to past gradients and thereby adapt to the geometry of the objective function. Variants such as RMSprop and Adam demonstrate outstanding practical performance that have contributed to the success of deep learning. In this work, we present AdaLVR, which combines the AdaGrad algorithm with loopless variance-reduced gradient estimators such as SAGA or L-SVRG that benefits from a straightforward construction and a streamlined analysis. We assess that AdaLVR inherits both good convergence properties from VR methods and the adaptive nature of AdaGrad: in the case of L-smooth convex functions we establish a gradient complexity of O(n + (L + √ nL)/ε) without prior knowledge of L. Numerical experiments demonstrate the superiority of AdaLVR over state-of-the-art methods. Moreover, we empirically show that the RMSprop and Adam algorithm combined with variance-reduced gradients estimators achieve even faster convergence.

Keywords: convex optimization, variance reduction, adaptive algorithms, loopless

Procedia PDF Downloads 57
808 An Energy Efficient Spectrum Shaping Scheme for Substrate Integrated Waveguides Based on Spread Reshaping Code

Authors: Yu Zhao, Rainer Gruenheid, Gerhard Bauch

Abstract:

In the microwave and millimeter-wave transmission region, substrate-integrated waveguide (SIW) is a very promising candidate for the development of circuits and components. It facilitates the transmission at the data rates in excess of 200 Gbit/s. An SIW mimics a rectangular waveguide by approximating the closed sidewalls with a via fence. This structure suppresses the low frequency components and makes the channel of the SIW a bandpass or high pass filter. This channel characteristic impedes the conventional baseband transmission using non-return-to-zero (NRZ) pulse shaping scheme. Therefore, mixers are commonly proposed to be used as carrier modulator and demodulator in order to facilitate a passband transmission. However, carrier modulation is not an energy efficient solution, because modulation and demodulation at high frequencies consume a lot of energy. For the first time to our knowledge, this paper proposes a spectrum shaping scheme of low complexity for the channel of SIW, namely spread reshaping code. It aims at matching the spectrum of the transmit signal to the channel frequency response. It facilitates the transmission through the SIW channel while it avoids using carrier modulation. In some cases, it even does not need equalization. Simulations reveal a good performance of this scheme, such that, as a result, eye opening is achieved without any equalization or modulation for the respective transmission channels.

Keywords: bandpass channel, eye-opening, switching frequency, substrate-integrated waveguide, spectrum shaping scheme, spread reshaping code

Procedia PDF Downloads 150
807 Using Genetic Algorithms to Outline Crop Rotations and a Cropping-System Model

Authors: Nicolae Bold, Daniel Nijloveanu

Abstract:

The idea of cropping-system is a method used by farmers. It is an environmentally-friendly method, protecting the natural resources (soil, water, air, nutritive substances) and increase the production at the same time, taking into account some crop particularities. The combination of this powerful method with the concepts of genetic algorithms results into a possibility of generating sequences of crops in order to form a rotation. The usage of this type of algorithms has been efficient in solving problems related to optimization and their polynomial complexity allows them to be used at solving more difficult and various problems. In our case, the optimization consists in finding the most profitable rotation of cultures. One of the expected results is to optimize the usage of the resources, in order to minimize the costs and maximize the profit. In order to achieve these goals, a genetic algorithm was designed. This algorithm ensures the finding of several optimized solutions of cropping-systems possibilities which have the highest profit and, thus, which minimize the costs. The algorithm uses genetic-based methods (mutation, crossover) and structures (genes, chromosomes). A cropping-system possibility will be considered a chromosome and a crop within the rotation is a gene within a chromosome. Results about the efficiency of this method will be presented in a special section. The implementation of this method would bring benefits into the activity of the farmers by giving them hints and helping them to use the resources efficiently.

Keywords: chromosomes, cropping, genetic algorithm, genes

Procedia PDF Downloads 417
806 Screening and Optimization of Pretreatments for Rice Straw and Their Utilization for Bioethanol Production Using Developed Yeast Strain

Authors: Ganesh Dattatraya Saratale, Min Kyu Oh

Abstract:

Rice straw is one of the most abundant lignocellulosic waste materials and its annual production is about 731 Mt in the world. This study treats the subject of effective utilization of this waste biomass for biofuels production. We have showed a comparative assessment of numerous pretreatment strategies for rice straw, comprising of major physical, chemical and physicochemical methods. Among the different methods employed for pretreatment alkaline pretreatment in combination with sodium chlorite/acetic acid delignification found efficient pretreatment with significant improvement in the enzymatic digestibility of rice straw. A cellulase dose of 20 filter paper units (FPU) released a maximum 63.21 g/L of reducing sugar with 94.45% hydrolysis yield and 64.64% glucose yield from rice straw, respectively. The effects of different pretreatment methods on biomass structure and complexity were investigated by FTIR, XRD and SEM analytical techniques. Finally the enzymatic hydrolysate of rice straw was used for ethanol production using developed Saccharomyces cerevisiae SR8. The developed yeast strain enabled efficient fermentation of xylose and glucose and produced higher ethanol production. Thus development of bioethanol production from lignocellulosic waste biomass is generic, applicable methodology and have great implication for using ‘green raw materials’ and producing ‘green products’ much needed today.

Keywords: rice straw, pretreatment, enzymatic hydrolysis, FPU, Saccharomyces cerevisiae SR8, ethanol fermentation

Procedia PDF Downloads 525
805 Power Iteration Clustering Based on Deflation Technique on Large Scale Graphs

Authors: Taysir Soliman

Abstract:

One of the current popular clustering techniques is Spectral Clustering (SC) because of its advantages over conventional approaches such as hierarchical clustering, k-means, etc. and other techniques as well. However, one of the disadvantages of SC is the time consuming process because it requires computing the eigenvectors. In the past to overcome this disadvantage, a number of attempts have been proposed such as the Power Iteration Clustering (PIC) technique, which is one of versions from SC; some of PIC advantages are: 1) its scalability and efficiency, 2) finding one pseudo-eigenvectors instead of computing eigenvectors, and 3) linear combination of the eigenvectors in linear time. However, its worst disadvantage is an inter-class collision problem because it used only one pseudo-eigenvectors which is not enough. Previous researchers developed Deflation-based Power Iteration Clustering (DPIC) to overcome problems of PIC technique on inter-class collision with the same efficiency of PIC. In this paper, we developed Parallel DPIC (PDPIC) to improve the time and memory complexity which is run on apache spark framework using sparse matrix. To test the performance of PDPIC, we compared it to SC, ESCG, ESCALG algorithms on four small graph benchmark datasets and nine large graph benchmark datasets, where PDPIC proved higher accuracy and better time consuming than other compared algorithms.

Keywords: spectral clustering, power iteration clustering, deflation-based power iteration clustering, Apache spark, large graph

Procedia PDF Downloads 178
804 GeoWeb at the Service of Household Waste Collection in Urban Areas

Authors: Abdessalam Hijab, Eric Henry, Hafida Boulekbache

Abstract:

The complexity of the city makes sustainable management of the urban environment more difficult. Managers are required to make significant human and technical investments, particularly in household waste collection (focus of our research). The aim of this communication is to propose a collaborative geographic multi-actor device (MGCD) based on the link between information and communication technologies (ICT) and geo-web tools in order to involve urban residents in household waste collection processes. Our method is based on a collaborative/motivational concept between the city and its residents. It is a geographic collaboration dedicated to the general public (citizens, residents, and any other participant), based on real-time allocation and geographic location of topological, geographic, and multimedia data in the form of local geo-alerts (location-specific problems) related to household waste in an urban environment. This contribution allows us to understand the extent to which residents can assist and contribute to the development of household waste collection processes for a better protected urban environment. This suggestion provides a good idea of how residents can contribute to the data bank for future uses. Moreover, it will contribute to the transformation of the population into a smart inhabitant as an essential component of a smart city. The proposed model will be tested in the Lamkansa sampling district in Casablanca, Morocco.

Keywords: information and communication technologies, ICTs, GeoWeb, geo-collaboration, city, inhabitant, waste, collection, environment

Procedia PDF Downloads 111
803 Workforce Optimization: Fair Workload Balance and Near-Optimal Task Execution Order

Authors: Alvaro Javier Ortega

Abstract:

A large number of companies face the challenge of matching highly-skilled professionals to high-end positions by human resource deployment professionals. However, when the professional list and tasks to be matched are larger than a few dozens, this process result is far from optimal and takes a long time to be made. Therefore, an automated assignment algorithm for this workforce management problem is needed. The majority of companies are divided into several sectors or departments, where trained employees with different experience levels deal with a large number of tasks daily. Also, the execution order of all tasks is of mater consequence, due to some of these tasks just can be run it if the result of another task is provided. Thus, a wrong execution order leads to large waiting times between consecutive tasks. The desired goal is, therefore, creating accurate matches and a near-optimal execution order that maximizes the number of tasks performed and minimizes the idle time of the expensive skilled employees. The problem described before can be model as a mixed-integer non-linear programming (MINLP) as it will be shown in detail through this paper. A large number of MINLP algorithms have been proposed in the literature. Here, genetic algorithm solutions are considered and a comparison between two different mutation approaches is presented. The simulated results considering different complexity levels of assignment decisions show the appropriateness of the proposed model.

Keywords: employees, genetic algorithm, industry management, workforce

Procedia PDF Downloads 158
802 Technology Transfer and FDI: Some Lessons for Tunisia

Authors: Assaad Ghazouani, Hedia Teraoui

Abstract:

The purpose of this article is to try to see if the FDI actually contributes to technology transfer in Tunisia or are there other sources that can guarantee this transfer? The answer to this problem was gradual as we followed an approach using economic theory, the reality of Tunisia and econometric and statistical tools. We examined the relationship between technology transfer and FDI in Tunisia over a period of 40 years from 1970 to 2010. We estimated in two stages: first, a growth equation, then we have learned from this regression residue (proxy technology), secondly, we regressed on European FDI, exports of manufactures, imports of goods from the European Union in addition to other variables to test the robustness of the results and describing the level of infrastructure in the country. It follows from our study that technology transfer does not originate primarily and exclusively in the FDI and the latter is econometrically weakly with technology transfer and spill over effect of FDI does not seem to occur according to our results. However, the relationship between technology transfer and imports is negative and significant. Although this result is cons-intuitive, is recurrent in the literature of panel data. It has also given rise to intense debate on the microeconomic modelling as well as on the empirical applications. Technology transfer through trade or foreign investment has become a catalyst for growth recognized by numerous empirical studies in particular. However, the relationship technology transfer FDI is more complex than it appears. This complexity is due, primarily, but not exclusively to the close link between FDI and the characteristics of the host country. This is essentially the host's responsibility to establish general conditions, transparent and conducive to investment, and to strengthen human and institutional capacity necessary for foreign capital flows that can have real effects on growth.

Keywords: technology transfer, foreign direct investment, economics, finance

Procedia PDF Downloads 312
801 Identification of Hepatocellular Carcinoma Using Supervised Learning Algorithms

Authors: Sagri Sharma

Abstract:

Analysis of diseases integrating multi-factors increases the complexity of the problem and therefore, development of frameworks for the analysis of diseases is an issue that is currently a topic of intense research. Due to the inter-dependence of the various parameters, the use of traditional methodologies has not been very effective. Consequently, newer methodologies are being sought to deal with the problem. Supervised Learning Algorithms are commonly used for performing the prediction on previously unseen data. These algorithms are commonly used for applications in fields ranging from image analysis to protein structure and function prediction and they get trained using a known dataset to come up with a predictor model that generates reasonable predictions for the response to new data. Gene expression profiles generated by DNA analysis experiments can be quite complex since these experiments can involve hypotheses involving entire genomes. The application of well-known machine learning algorithm - Support Vector Machine - to analyze the expression levels of thousands of genes simultaneously in a timely, automated and cost effective way is thus used. The objectives to undertake the presented work are development of a methodology to identify genes relevant to Hepatocellular Carcinoma (HCC) from gene expression dataset utilizing supervised learning algorithms and statistical evaluations along with development of a predictive framework that can perform classification tasks on new, unseen data.

Keywords: artificial intelligence, biomarker, gene expression datasets, hepatocellular carcinoma, machine learning, supervised learning algorithms, support vector machine

Procedia PDF Downloads 419
800 Advanced Combinatorial Method for Solving Complex Fault Trees

Authors: José de Jesús Rivero Oliva, Jesús Salomón Llanes, Manuel Perdomo Ojeda, Antonio Torres Valle

Abstract:

Combinatorial explosion is a common problem to both predominant methods for solving fault trees: Minimal Cut Set (MCS) approach and Binary Decision Diagram (BDD). High memory consumption impedes the complete solution of very complex fault trees. Only approximated non-conservative solutions are possible in these cases using truncation or other simplification techniques. The paper proposes a method (CSolv+) for solving complex fault trees, without any possibility of combinatorial explosion. Each individual MCS is immediately discarded after its contribution to the basic events importance measures and the Top gate Upper Bound Probability (TUBP) has been accounted. An estimation of the Top gate Exact Probability (TEP) is also provided. Therefore, running in a computer cluster, CSolv+ will guarantee the complete solution of complex fault trees. It was successfully applied to 40 fault trees from the Aralia fault trees database, performing the evaluation of the top gate probability, the 1000 Significant MCSs (SMCS), and the Fussell-Vesely, RRW and RAW importance measures for all basic events. The high complexity fault tree nus9601 was solved with truncation probabilities from 10-²¹ to 10-²⁷ just to limit the execution time. The solution corresponding to 10-²⁷ evaluated 3.530.592.796 MCSs in 3 hours and 15 minutes.

Keywords: system reliability analysis, probabilistic risk assessment, fault tree analysis, basic events importance measures

Procedia PDF Downloads 28