Search results for: complex coacervation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5075

Search results for: complex coacervation

4235 Bayesian Parameter Inference for Continuous Time Markov Chains with Intractable Likelihood

Authors: Randa Alharbi, Vladislav Vyshemirsky

Abstract:

Systems biology is an important field in science which focuses on studying behaviour of biological systems. Modelling is required to produce detailed description of the elements of a biological system, their function, and their interactions. A well-designed model requires selecting a suitable mechanism which can capture the main features of the system, define the essential components of the system and represent an appropriate law that can define the interactions between its components. Complex biological systems exhibit stochastic behaviour. Thus, using probabilistic models are suitable to describe and analyse biological systems. Continuous-Time Markov Chain (CTMC) is one of the probabilistic models that describe the system as a set of discrete states with continuous time transitions between them. The system is then characterised by a set of probability distributions that describe the transition from one state to another at a given time. The evolution of these probabilities through time can be obtained by chemical master equation which is analytically intractable but it can be simulated. Uncertain parameters of such a model can be inferred using methods of Bayesian inference. Yet, inference in such a complex system is challenging as it requires the evaluation of the likelihood which is intractable in most cases. There are different statistical methods that allow simulating from the model despite intractability of the likelihood. Approximate Bayesian computation is a common approach for tackling inference which relies on simulation of the model to approximate the intractable likelihood. Particle Markov chain Monte Carlo (PMCMC) is another approach which is based on using sequential Monte Carlo to estimate intractable likelihood. However, both methods are computationally expensive. In this paper we discuss the efficiency and possible practical issues for each method, taking into account the computational time for these methods. We demonstrate likelihood-free inference by performing analysing a model of the Repressilator using both methods. Detailed investigation is performed to quantify the difference between these methods in terms of efficiency and computational cost.

Keywords: Approximate Bayesian computation(ABC), Continuous-Time Markov Chains, Sequential Monte Carlo, Particle Markov chain Monte Carlo (PMCMC)

Procedia PDF Downloads 195
4234 Toward a Coalitional Subject in Contemporary American Feminist Literature

Authors: Su-Lin Yu

Abstract:

Coalition politics has been one of feminists’ persistent concerns. Following recent feminist discussion on new modes of affiliation across difference, she will explore how the process of female subject formation depends on alliances across different cultural locations. First, she will examine how coalition politics is reformulated across difference in contemporary feminist literature. In particular, the paper will identify the particular contexts and locations in which coalition building both enables and constrains the female subject. She will attempt to explore how contemporary feminist literature highlights the possibilities and limitations for solidarity and affiliations. To understand coalition politics in contemporary feminist works, she will engage in close readings of two texts: Rebecca Walker’s Black, White and Jewish: Memoir of a Shifting Self and Danzy Senna’s Caucasia. Both Walker and Senna have articulated the complex nodes of identity that are staged by a politics of location as they refuse to be boxed into simplistic essentialist positions. Their texts are characterized by the characters’ racial ambiguity and their social and geographical mobility of life in the contemporary United States. Their experiences of living through conflictual and contradictory relationships never fully fit the boundaries of racial categorization. Each of these texts demonstrates the limits as well as the possibilities of working with diversity among and within persons and groups, thus, laying the ground for complex alliance formation. Because each of the protagonists must negotiate a set of contradictions, they will have to constantly shift their affiliations. Rather than construct a static alliance, they describe a process of moving ‘beyond boundaries,’ an embracing of multiple locations. As self-identified third wavers, Rebecca Walker and Danzy Senna have been identified and marked with the status of ‘leader’ by the feminist establishment and by mainstream U.S. media. Their texts have captured both mass popularity and critical attention in the feminist and, often, the non-feminist literary community. By analyzing these texts, she will show how contemporary American feminist literature reveals coalition politics which is fraught with complications and unintended consequences. Taken as a whole, then, these works provide an important examination not only of coalition politics of American feminism, but also a snapshot of a central debate among feminist critique of coalition politics as a whole.

Keywords: coalition politics, contemporary women’s literature, identity, female subject

Procedia PDF Downloads 279
4233 Site Specific Nutrient Management Need in India Now

Authors: A. H. Nanher, N. P. Singh, Shashidhar Yadav, Sachin Tyagi

Abstract:

Agricultural production system is an outcome of a complex interaction of seed, soil, water and agro-chemicals (including fertilizers). Therefore, judicious management of all the inputs is essential for the sustainability of such a complex system. Precision agriculture gives farmers the ability to use crop inputs more effectively including fertilizers, pesticides, tillage and irrigation water. More effective use of inputs means greater crop yield and/or quality, without polluting the environment the focus on enhancing the productivity during the Green Revolution coupled with total disregard of proper management of inputs and without considering the ecological impacts, has resulted into environmental degradation. To evaluate a new approach for site-specific nutrient management (SSNM). Large variation in initial soil fertility characteristics and indigenous supply of N, P, and K was observed among Field- and season-specific NPK applications were calculated by accounting for the indigenous nutrient supply, yield targets, and nutrient demand as a function of the interactions between N, P, and K. Nitrogen applications were fine-tuned based on season-specific rules and field-specific monitoring of crop N status. The performance of SSNM did not differ significantly between high-yielding and low-yielding climatic seasons, but improved over time with larger benefits observed in the second year Future, strategies for nutrient management in intensive rice systems must become more site-specific and dynamic to manage spatially and temporally variable resources based on a quantitative understanding of the congruence between nutrient supply and crop demand. The SSNM concept has demonstrated promising agronomic and economic potential. It can be used for managing plant nutrients at any scale, i.e., ranging from a general recommendation for homogenous management of a larger domain to true management of between-field variability. Assessment of pest profiles in FFP and SSNM plots suggests that SSNM may also reduce pest incidence, particularly diseases that are often associated with excessive N use or unbalanced plant nutrition.

Keywords: nutrient, pesticide, crop, yield

Procedia PDF Downloads 417
4232 Phenotype Prediction of DNA Sequence Data: A Machine and Statistical Learning Approach

Authors: Mpho Mokoatle, Darlington Mapiye, James Mashiyane, Stephanie Muller, Gciniwe Dlamini

Abstract:

Great advances in high-throughput sequencing technologies have resulted in availability of huge amounts of sequencing data in public and private repositories, enabling a holistic understanding of complex biological phenomena. Sequence data are used for a wide range of applications such as gene annotations, expression studies, personalized treatment and precision medicine. However, this rapid growth in sequence data poses a great challenge which calls for novel data processing and analytic methods, as well as huge computing resources. In this work, a machine and statistical learning approach for DNA sequence classification based on $k$-mer representation of sequence data is proposed. The approach is tested using whole genome sequences of Mycobacterium tuberculosis (MTB) isolates to (i) reduce the size of genomic sequence data, (ii) identify an optimum size of k-mers and utilize it to build classification models, (iii) predict the phenotype from whole genome sequence data of a given bacterial isolate, and (iv) demonstrate computing challenges associated with the analysis of whole genome sequence data in producing interpretable and explainable insights. The classification models were trained on 104 whole genome sequences of MTB isoloates. Cluster analysis showed that k-mers maybe used to discriminate phenotypes and the discrimination becomes more concise as the size of k-mers increase. The best performing classification model had a k-mer size of 10 (longest k-mer) an accuracy, recall, precision, specificity, and Matthews Correlation coeffient of 72.0%, 80.5%, 80.5%, 63.6%, and 0.4 respectively. This study provides a comprehensive approach for resampling whole genome sequencing data, objectively selecting a k-mer size, and performing classification for phenotype prediction. The analysis also highlights the importance of increasing the k-mer size to produce more biological explainable results, which brings to the fore the interplay that exists amongst accuracy, computing resources and explainability of classification results. However, the analysis provides a new way to elucidate genetic information from genomic data, and identify phenotype relationships which are important especially in explaining complex biological mechanisms.

Keywords: AWD-LSTM, bootstrapping, k-mers, next generation sequencing

Procedia PDF Downloads 156
4231 Phenotype Prediction of DNA Sequence Data: A Machine and Statistical Learning Approach

Authors: Darlington Mapiye, Mpho Mokoatle, James Mashiyane, Stephanie Muller, Gciniwe Dlamini

Abstract:

Great advances in high-throughput sequencing technologies have resulted in availability of huge amounts of sequencing data in public and private repositories, enabling a holistic understanding of complex biological phenomena. Sequence data are used for a wide range of applications such as gene annotations, expression studies, personalized treatment and precision medicine. However, this rapid growth in sequence data poses a great challenge which calls for novel data processing and analytic methods, as well as huge computing resources. In this work, a machine and statistical learning approach for DNA sequence classification based on k-mer representation of sequence data is proposed. The approach is tested using whole genome sequences of Mycobacterium tuberculosis (MTB) isolates to (i) reduce the size of genomic sequence data, (ii) identify an optimum size of k-mers and utilize it to build classification models, (iii) predict the phenotype from whole genome sequence data of a given bacterial isolate, and (iv) demonstrate computing challenges associated with the analysis of whole genome sequence data in producing interpretable and explainable insights. The classification models were trained on 104 whole genome sequences of MTB isoloates. Cluster analysis showed that k-mers maybe used to discriminate phenotypes and the discrimination becomes more concise as the size of k-mers increase. The best performing classification model had a k-mer size of 10 (longest k-mer) an accuracy, recall, precision, specificity, and Matthews Correlation coeffient of 72.0 %, 80.5 %, 80.5 %, 63.6 %, and 0.4 respectively. This study provides a comprehensive approach for resampling whole genome sequencing data, objectively selecting a k-mer size, and performing classification for phenotype prediction. The analysis also highlights the importance of increasing the k-mer size to produce more biological explainable results, which brings to the fore the interplay that exists amongst accuracy, computing resources and explainability of classification results. However, the analysis provides a new way to elucidate genetic information from genomic data, and identify phenotype relationships which are important especially in explaining complex biological mechanisms

Keywords: AWD-LSTM, bootstrapping, k-mers, next generation sequencing

Procedia PDF Downloads 142
4230 Mesoporous Na2Ti3O7 Nanotube-Constructed Materials with Hierarchical Architecture: Synthesis and Properties

Authors: Neumoin Anton Ivanovich, Opra Denis Pavlovich

Abstract:

Materials based on titanium oxide compounds are widely used in such areas as solar energy, photocatalysis, food industry and hygiene products, biomedical technologies, etc. Demand for them has also formed in the battery industry (an example of this is the commercialization of Li4Ti5O12), where much attention has recently been paid to the development of next-generation systems and technologies, such as sodium-ion batteries. This dictates the need to search for new materials with improved characteristics, as well as ways to obtain them that meet the requirements of scalability. One of the ways to solve these problems can be the creation of nanomaterials that often have a complex of physicochemical properties that radically differ from the characteristics of their counterparts in the micro- or macroscopic state. At the same time, it is important to control the texture (specific surface area, porosity) of such materials. In view of the above, among other methods, the hydrothermal technique seems to be suitable, allowing a wide range of control over the conditions of synthesis. In the present study, a method was developed for the preparation of mesoporous nanostructured sodium trititanate (Na2Ti3O7) with a hierarchical architecture. The materials were synthesized by hydrothermal processing and exhibit a complex hierarchically organized two-layer architecture. At the first level of the hierarchy, materials are represented by particles having a roughness surface, and at the second level, by one-dimensional nanotubes. The products were found to have high specific surface area and porosity with a narrow pore size distribution (about 6 nm). As it is known, the specific surface area and porosity are important characteristics of functional materials, which largely determine the possibilities and directions of their practical application. Electrochemical impedance spectroscopy data show that the resulting sodium trititanate has a sufficiently high electrical conductivity. As expected, the synthesized complexly organized nanoarchitecture based on sodium trititanate with a porous structure can be practically in demand, for example, in the field of new generation electrochemical storage and energy conversion devices.

Keywords: sodium trititanate, hierarchical materials, mesoporosity, nanotubes, hydrothermal synthesis

Procedia PDF Downloads 97
4229 Robust Control of a Parallel 3-RRR Robotic Manipulator via μ-Synthesis Method

Authors: A. Abbasi Moshaii, M. Soltan Rezaee, M. Mohammadi Moghaddam

Abstract:

Control of some mechanisms is hard because of their complex dynamic equations. If part of the complexity is resulting from uncertainties, an efficient way for solving that is robust control. By this way, the control procedure could be simple and fast and finally, a simple controller can be designed. One kind of these mechanisms is 3-RRR which is a parallel mechanism and has three revolute joints. This paper aims to robust control a 3-RRR planner mechanism and it presents that this could be used for other mechanisms. So, a significant problem in mechanisms control could be solved. The relevant diagrams are drawn and they show the correctness of control process.

Keywords: 3-RRR, dynamic equations, mechanisms control, structural uncertainty

Procedia PDF Downloads 544
4228 Spatio-Temporal Analysis of Land Use Change and Green Cover Index

Authors: Poonam Sharma, Ankur Srivastav

Abstract:

Cities are complex and dynamic systems that constitute a significant challenge to urban planning. The increasing size of the built-up area owing to growing population pressure and economic growth have lead to massive Landuse/Landcover change resulted in the loss of natural habitat and thus reducing the green covers in urban areas. Urban environmental quality is influenced by several aspects, including its geographical configuration, the scale, and nature of human activities occurring and environmental impacts generated. Cities have transformed into complex and dynamic systems that constitute a significant challenge to urban planning. Cities and their sustainability are often discussed together as the cities stand confronted with numerous environmental concerns as the world becoming increasingly urbanized, and the cities are situated in the mesh of global networks in multiple senses. A rapid transformed urban setting plays a crucial role to change the green area of natural habitats. To examine the pattern of urban growth and to measure the Landuse/Landcover change in Gurgoan in Haryana, India through the integration of Geospatial technique is attempted in the research paper. Satellite images are used to measure the spatiotemporal changes that have occurred in the land use and land cover resulting into a new cityscape. It has been observed from the analysis that drastically evident changes in land use has occurred with the massive rise in built up areas and the decrease in green cover and therefore causing the sustainability of the city an important area of concern. The massive increase in built-up area has influenced the localised temperatures and heat concentration. To enhance the decision-making process in urban planning, a detailed and real world depiction of these urban spaces is the need of the hour. Monitoring indicators of key processes in land use and economic development are essential for evaluating policy measures.

Keywords: cityscape, geospatial techniques, green cover index, urban environmental quality, urban planning

Procedia PDF Downloads 257
4227 Development of Building Information Modeling in Property Industry: Beginning with Building Information Modeling Construction

Authors: B. Godefroy, D. Beladjine, K. Beddiar

Abstract:

In France, construction BIM actors commonly evoke the BIM gains for exploitation by integrating of the life cycle of a building. The standardization of level 7 of development would achieve this stage of the digital model. The householders include local public authorities, social landlords, public institutions (health and education), enterprises, facilities management companies. They have a dual role: owner and manager of their housing complex. In a context of financial constraint, the BIM of exploitation aims to control costs, make long-term investment choices, renew the portfolio and enable environmental standards to be met. It assumes a knowledge of the existing buildings, marked by its size and complexity. The information sought must be synthetic and structured, it concerns, in general, a real estate complex. We conducted a study with professionals about their concerns and ways to use it to see how householders could benefit from this development. To obtain results, we had in mind the recurring interrogation of the project management, on the needs of the operators, we tested the following stages: 1) Inculcate a minimal culture of BIM with multidisciplinary teams of the operator then by business, 2) Learn by BIM tools, the adaptation of their trade in operations, 3) Understand the place and creation of a graphic and technical database management system, determine the components of its library so their needs, 4) Identify the cross-functional interventions of its managers by business (operations, technical, information system, purchasing and legal aspects), 5) Set an internal protocol and define the BIM impact in their digital strategy. In addition, continuity of management by the integration of construction models in the operation phase raises the question of interoperability in the control of the production of IFC files in the operator’s proprietary format and the export and import processes, a solution rivaled by the traditional method of vectorization of paper plans. Companies that digitize housing complex and those in FM produce a file IFC, directly, according to their needs without recourse to the model of construction, they produce models business for the exploitation. They standardize components, equipment that are useful for coding. We observed the consequences resulting from the use of the BIM in the property industry and, made the following observations: a) The value of data prevail over the graphics, 3D is little used b) The owner must, through his organization, promote the feedback of technical management information during the design phase c) The operator's reflection on outsourcing concerns the acquisition of its information system and these services, observing the risks and costs related to their internal or external developments. This study allows us to highlight: i) The need for an internal organization of operators prior to a response to the construction management ii) The evolution towards automated methods for creating models dedicated to the exploitation, a specialization would be required iii) A review of the communication of the project management, management continuity not articulating around his building model, it must take into account the environment of the operator and reflect on its scope of action.

Keywords: information system, interoperability, models for exploitation, property industry

Procedia PDF Downloads 134
4226 Roles of Tester in Automated World

Authors: Sagar Mahendrakar

Abstract:

Testers' roles have changed dramatically as automation continues to revolutionise the software development lifecycle. There's a general belief that manual testing is becoming outdated with the introduction of advanced testing frameworks and tools. This abstract, however, disproves that notion by examining the complex and dynamic role that testers play in automated environments. In this work, we explore the complex duties that testers have when everything is automated. We contend that although automation increases productivity and simplifies monotonous tasks, it cannot completely replace the cognitive abilities and subject-matter knowledge of human testers. Rather, testers shift their focus to higher-value tasks like creating test strategies, designing test cases, and delving into intricate scenarios that are difficult to automate. We also emphasise the critical role that testers play in guaranteeing the precision, thoroughness, and dependability of automated testing. Testers verify the efficacy of automated scripts and pinpoint areas for improvement through rigorous test planning, execution, and result analysis. They play the role of quality defenders, using their analytical and problem-solving abilities to find minute flaws that computerised tests might miss. Furthermore, the abstract emphasises how testing in automated environments is a collaborative process. In order to match testing efforts with business objectives, improve test automation frameworks, and rank testing tasks according to risk, testers work closely with developers, automation engineers, and other stakeholders. Finally, we discuss how testers in the era of automation need to possess a growing skill set. To stay current, testers need to develop skills in scripting languages, test automation tools, and emerging technologies in addition to traditional testing competencies. Soft skills like teamwork, communication, and flexibility are also essential for productive cooperation in cross-functional teams. This abstract clarifies the ongoing importance of testers in automated settings. Testers can use automation to improve software quality and provide outstanding user experiences by accepting their changing role as strategic partners and advocates for quality.

Keywords: testing, QA, automation, leadership

Procedia PDF Downloads 28
4225 Intersectionality and Sensemaking: Advancing the Conversation on Leadership as the Management of Meaning

Authors: Clifford Lewis

Abstract:

This paper aims to advance the conversation of an alternative view of leadership, namely ‘leadership as the management of meaning’. Here, leadership is considered as a social process of the management of meaning within an employment context, as opposed to a psychological trait, set of behaviours or relational consequence as seen in mainstream leadership research. Specifically, this study explores the relationship between intersectional identities and the management of meaning. Design: Semi-structured, one-on-one interviews were conducted with women and men of colour working in the South African private sector organisations in various leadership positions. Employing an intersectional approach using gender and race, participants were selected by using purposive and snowball sampling concurrently. Thematic and Axial coding was used to identify dominant themes. Findings: Findings suggest that, both gender and race shape how leaders manage meaning. Findings also confirm that intersectionality is an appropriate approach when studying the leadership experiences of those groups who are underrepresented in organisational leadership structures. The findings points to the need for further research into the differential effects of intersecting identities on organisational leadership experiences and that ‘leadership as the management of meaning’ is an appropriate approach for addressing this knowledge gap. Theoretical Contribution: There is a large body of literature on the complex challenges faced by women and people of colour in leadership but there is relatively little empirical work on how identity influences the management of meaning. This study contributes to the leadership literature by providing insight into how intersectional identities influence the management of meaning at work and how this impacts the leadership experiences of largely marginalised groups. Practical Implications: Understanding the leadership experiences of underrepresented groups is important because of both legal mandates and for building diverse talent for organisations and societies. Such an understanding assists practitioners in being sensitive to simplistic notions of challenges individuals might face in accessing and practicing leadership in organisations. Advancing the conversation on leadership as the management of meaning allows for a better understanding of complex challenges faced by women and people of colour and an opportunity for organisations to systematically remove unfair structural obstacles and develop their diverse leadership capacity.

Keywords: intersectionality, diversity, leadership, sensemaking

Procedia PDF Downloads 260
4224 A Resilience-Based Approach for Assessing Social Vulnerability in New Zealand's Coastal Areas

Authors: Javad Jozaei, Rob G. Bell, Paula Blackett, Scott A. Stephens

Abstract:

In the last few decades, Social Vulnerability Assessment (SVA) has been a favoured means in evaluating the susceptibility of social systems to drivers of change, including climate change and natural disasters. However, the application of SVA to inform responsive and practical strategies to deal with uncertain climate change impacts has always been challenging, and typically agencies resort back to conventional risk/vulnerability assessment. These challenges include complex nature of social vulnerability concepts which influence its applicability, complications in identifying and measuring social vulnerability determinants, the transitory social dynamics in a changing environment, and unpredictability of the scenarios of change that impacts the regime of vulnerability (including contention of when these impacts might emerge). Research suggests that the conventional quantitative approaches in SVA could not appropriately address these problems; hence, the outcomes could potentially be misleading and not fit for addressing the ongoing uncertain rise in risk. The second phase of New Zealand’s Resilience to Nature’s Challenges (RNC2) is developing a forward-looking vulnerability assessment framework and methodology that informs the decision-making and policy development in dealing with the changing coastal systems and accounts for complex dynamics of New Zealand’s coastal systems (including socio-economic, environmental and cultural). Also, RNC2 requires the new methodology to consider plausible drivers of incremental and unknowable changes, create mechanisms to enhance social and community resilience; and fits the New Zealand’s multi-layer governance system. This paper aims to analyse the conventional approaches and methodologies in SVA and offer recommendations for more responsive approaches that inform adaptive decision-making and policy development in practice. The research adopts a qualitative research design to examine different aspects of the conventional SVA processes, and the methods to achieve the research objectives include a systematic review of the literature and case study methods. We found that the conventional quantitative, reductionist and deterministic mindset in the SVA processes -with a focus the impacts of rapid stressors (i.e. tsunamis, floods)- show some deficiencies to account for complex dynamics of social-ecological systems (SES), and the uncertain, long-term impacts of incremental drivers. The paper will focus on addressing the links between resilience and vulnerability; and suggests how resilience theory and its underpinning notions such as the adaptive cycle, panarchy, and system transformability could address these issues, therefore, influence the perception of vulnerability regime and its assessment processes. In this regard, it will be argued that how a shift of paradigm from ‘specific resilience’, which focuses on adaptive capacity associated with the notion of ‘bouncing back’, to ‘general resilience’, which accounts for system transformability, regime shift, ‘bouncing forward’, can deliver more effective strategies in an era characterised by ongoing change and deep uncertainty.

Keywords: complexity, social vulnerability, resilience, transformation, uncertain risks

Procedia PDF Downloads 83
4223 Comparative Analysis of Reinforcement Learning Algorithms for Autonomous Driving

Authors: Migena Mana, Ahmed Khalid Syed, Abdul Malik, Nikhil Cherian

Abstract:

In recent years, advancements in deep learning enabled researchers to tackle the problem of self-driving cars. Car companies use huge datasets to train their deep learning models to make autonomous cars a reality. However, this approach has certain drawbacks in that the state space of possible actions for a car is so huge that there cannot be a dataset for every possible road scenario. To overcome this problem, the concept of reinforcement learning (RL) is being investigated in this research. Since the problem of autonomous driving can be modeled in a simulation, it lends itself naturally to the domain of reinforcement learning. The advantage of this approach is that we can model different and complex road scenarios in a simulation without having to deploy in the real world. The autonomous agent can learn to drive by finding the optimal policy. This learned model can then be easily deployed in a real-world setting. In this project, we focus on three RL algorithms: Q-learning, Deep Deterministic Policy Gradient (DDPG), and Proximal Policy Optimization (PPO). To model the environment, we have used TORCS (The Open Racing Car Simulator), which provides us with a strong foundation to test our model. The inputs to the algorithms are the sensor data provided by the simulator such as velocity, distance from side pavement, etc. The outcome of this research project is a comparative analysis of these algorithms. Based on the comparison, the PPO algorithm gives the best results. When using PPO algorithm, the reward is greater, and the acceleration, steering angle and braking are more stable compared to the other algorithms, which means that the agent learns to drive in a better and more efficient way in this case. Additionally, we have come up with a dataset taken from the training of the agent with DDPG and PPO algorithms. It contains all the steps of the agent during one full training in the form: (all input values, acceleration, steering angle, break, loss, reward). This study can serve as a base for further complex road scenarios. Furthermore, it can be enlarged in the field of computer vision, using the images to find the best policy.

Keywords: autonomous driving, DDPG (deep deterministic policy gradient), PPO (proximal policy optimization), reinforcement learning

Procedia PDF Downloads 134
4222 Supramolecular Approach towards Novel Applications: Battery, Band Gap and Gas Separation

Authors: Sudhakara Naidu Neppalli, Tejas S. Bhosale

Abstract:

It is well known that the block copolymer (BCP) can form a complex molecule, through non-covalent bonds such as hydrogen bond, ionic bond and co-ordination bond, with low molecular weight compound as well as with macromolecules, which provide vast applications, includes the alteration of morphology and properties of polymers. Hence we covered the research that, the importance of non-covalent bonds in increasing the non-favourable segmental interactions of the blocks was well examined by attaching and detaching the bonds between the BCP and additive. We also monitored the phase transition of block copolymer and effective interaction parameter (χeff) for Li-doped polymers using small angle x-ray scattering and transmission electron microscopy. The effective interaction parameter (χeff) between two block components was evaluated using Leibler theory based on the incompressible random phase approximation (RPA) for ionized BCP in a disordered state. Furthermore, conductivity experiments demonstrate that the ionic conductivity in the samples quenched from the different structures is morphology-independent, while it increases with increasing ion salt concentration. Morphological transitions, interaction parameter, and thermal stability also examined in quarternized block copolymer. D-spacing was used to estimate effective interaction parameter (χeff) of block components in weak and strong segregation regimes of ordered phase. Metal-containing polymer has been the topic of great attention in recent years due to their wide range of potential application. Similarly, metal- ligand complex is used as a supramolecular linker between the polymers giving rise to a ‘Metallo-Supramolecule assembly. More precisely, functionalized polymer end capped with 2, 2’:6’, 2”- terpyridine ligand can be selectively complexed with wide range of transition metal ions and then subsequently attached to other terpyridine terminated polymer block. In compare to other supramolecular assembly, BCP involved metallo-supramolecule assembly offers vast applications such as optical activity, electrical conductivity, luminescence and photo refractivity.

Keywords: band gap, block copolymer, conductivity, interaction parameter, phase transition

Procedia PDF Downloads 154
4221 Synchronization of Semiconductor Laser Networks

Authors: R. M. López-Gutiérrez, L. Cardoza-Avendaño, H. Cervantes-de Ávila, J. A. Michel-Macarty, C. Cruz-Hernández, A. Arellano-Delgado, R. Carmona-Rodríguez

Abstract:

In this paper, synchronization of multiple chaotic semiconductor lasers is achieved by appealing to complex system theory. In particular, we consider dynamical networks composed by semiconductor laser, as interconnected nodes, where the interaction in the networks are defined by coupling the first state of each node. An interesting case is synchronized with master-slave configuration in star topology. Nodes of these networks are modeled for the laser and simulated by Matlab. These results are applicable to private communication.

Keywords: chaotic laser, network, star topology, synchronization

Procedia PDF Downloads 549
4220 Building Envelope Engineering and Typologies for Complex Architectures: Composition and Functional Methodologies

Authors: Massimiliano Nastri

Abstract:

The study examines the façade systems according to the constitutive and typological characters, as well as the functional and applicative requirements such as the expressive, constructive, and interactive criteria towards the environmental, perceptive, and energy conditions. The envelope systems are understood as instruments of mediation, interchange, and dynamic interaction between environmental conditions. The façades are observed for the sustainable concept of eco-efficient envelopes, selective and multi-purpose filters, adaptable and adjustable according to the environmental performance.

Keywords: typologies of façades, environmental and energy sustainability, interaction and perceptive mediation, technical skins

Procedia PDF Downloads 138
4219 Impacts on Marine Ecosystems Using a Multilayer Network Approach

Authors: Nelson F. F. Ebecken, Gilberto C. Pereira, Lucio P. de Andrade

Abstract:

Bays, estuaries and coastal ecosystems are some of the most used and threatened natural systems globally. Its deterioration is due to intense and increasing human activities. This paper aims to monitor the socio-ecological in Brazil, model and simulate it through a multilayer network representing a DPSIR structure (Drivers, Pressures, States-Impacts-Responses) considering the concept of Management based on Ecosystems to support decision-making under the National/State/Municipal Coastal Management policy. This approach considers several interferences and can represent a significant advance in several scientific aspects. The main objective of this paper is the coupling of three different types of complex networks, the first being an ecological network, the second a social network, and the third a network of economic activities, in order to model the marine ecosystem. Multilayer networks comprise two or more "layers", which may represent different types of interactions, different communities, different points in time, and so on. The dependency between layers results from processes that affect the various layers. For example, the dispersion of individuals between two patches affects the network structure of both samples. A multilayer network consists of (i) a set of physical nodes representing entities (e.g., species, people, companies); (ii) a set of layers, which may include multiple layering aspects (e.g., time dependency and multiple types of relationships); (iii) a set of state nodes, each of which corresponds to the manifestation of a given physical node in a layer-specific; and (iv) a set of edges (weighted or not) to connect the state nodes among themselves. The edge set includes the intralayer edges familiar and interlayer ones, which connect state nodes between layers. The applied methodology in an existent case uses the Flow cytometry process and the modeling of ecological relationships (trophic and non-trophic) following fuzzy theory concepts and graph visualization. The identification of subnetworks in the fuzzy graphs is carried out using a specific computational method. This methodology allows considering the influence of different factors and helps their contributions to the decision-making process.

Keywords: marine ecosystems, complex systems, multilayer network, ecosystems management

Procedia PDF Downloads 99
4218 On the Zeros of the Degree Polynomial of a Graph

Authors: S. R. Nayaka, Putta Swamy

Abstract:

Graph polynomial is one of the algebraic representations of the Graph. The degree polynomial is one of the simple algebraic representations of graphs. The degree polynomial of a graph G of order n is the polynomial Deg(G, x) with the coefficients deg(G,i) where deg(G,i) denotes the number of vertices of degree i in G. In this article, we investigate the behavior of the roots of some families of Graphs in the complex field. We investigate for the graphs having only integral roots. Further, we characterize the graphs having single roots or having real roots and behavior of the polynomial at the particular value is also obtained.

Keywords: degree polynomial, regular graph, minimum and maximum degree, graph operations

Procedia PDF Downloads 234
4217 Advancements in Mathematical Modeling and Optimization for Control, Signal Processing, and Energy Systems

Authors: Zahid Ullah, Atlas Khan

Abstract:

This abstract focuses on the advancements in mathematical modeling and optimization techniques that play a crucial role in enhancing the efficiency, reliability, and performance of these systems. In this era of rapidly evolving technology, mathematical modeling and optimization offer powerful tools to tackle the complex challenges faced by control, signal processing, and energy systems. This abstract presents the latest research and developments in mathematical methodologies, encompassing areas such as control theory, system identification, signal processing algorithms, and energy optimization. The abstract highlights the interdisciplinary nature of mathematical modeling and optimization, showcasing their applications in a wide range of domains, including power systems, communication networks, industrial automation, and renewable energy. It explores key mathematical techniques, such as linear and nonlinear programming, convex optimization, stochastic modeling, and numerical algorithms, that enable the design, analysis, and optimization of complex control and signal processing systems. Furthermore, the abstract emphasizes the importance of addressing real-world challenges in control, signal processing, and energy systems through innovative mathematical approaches. It discusses the integration of mathematical models with data-driven approaches, machine learning, and artificial intelligence to enhance system performance, adaptability, and decision-making capabilities. The abstract also underscores the significance of bridging the gap between theoretical advancements and practical applications. It recognizes the need for practical implementation of mathematical models and optimization algorithms in real-world systems, considering factors such as scalability, computational efficiency, and robustness. In summary, this abstract showcases the advancements in mathematical modeling and optimization techniques for control, signal processing, and energy systems. It highlights the interdisciplinary nature of these techniques, their applications across various domains, and their potential to address real-world challenges. The abstract emphasizes the importance of practical implementation and integration with emerging technologies to drive innovation and improve the performance of control, signal processing, and energy.

Keywords: mathematical modeling, optimization, control systems, signal processing, energy systems, interdisciplinary applications, system identification, numerical algorithms

Procedia PDF Downloads 100
4216 An Alternative Way to Mapping Cone

Authors: Yousuf Alkhezi

Abstract:

Since most of the literature on algebra does not make much deal with the special case of mapping cone. This paper is an alternative way to examine the special tensor product and mapping cone. Also, we show that the isomorphism that implies the mapping cone commutes with the tensor product for the ordinary tensor product no longer holds for the pinched tensor product. However, we show there is a morphism. We will introduce an alternative way of mapping cone. We are looking for more properties which is our future project. Also, we want to apply these new properties in some application. Many results and examples with classical algorithms will be provided.

Keywords: complex, tensor product, pinched tensore product, mapping cone

Procedia PDF Downloads 118
4215 Development of Immersive Virtual Reality System for Planning of Cargo Loading Operations

Authors: Eugene Y. C. Wong, Daniel Y. W. Mo, Cosmo T. Y. Ng, Jessica K. Y. Chan, Leith K. Y. Chan, Henry Y. K. Lau

Abstract:

The real-time planning visualisation, precise allocation and loading optimisation in air cargo load planning operations are increasingly important as more considerations are needed on dangerous cargo loading, locations of lithium batteries, weight declaration and limited aircraft capacity. The planning of the unit load devices (ULD) can often be carried out only in a limited number of hours before flight departure. A dynamic air cargo load planning system is proposed with the optimisation of cargo load plan and visualisation of planning results in virtual reality systems. The system aims to optimise the cargo load planning and visualise the simulated loading planning decision on air cargo terminal operations. Adopting simulation tools, Cave Automatic Virtual Environment (CAVE) and virtual reality technologies, the results of planning with reference to weight and balance, Unit Load Device (ULD) dimensions, gateway, cargo nature and aircraft capacity are optimised and presented. The virtual reality system facilities planning, operations, education and training. Staff in terminals are usually trained in a traditional push-approach demonstration with enormous manual paperwork. With the support of newly customized immersive visualization environment, users can master the complex air cargo load planning techniques in a problem based training with the instant result being immersively visualised. The virtual reality system is developed with three-dimensional (3D) projectors, screens, workstations, truss system, 3D glasses, and demonstration platform and software. The content will be focused on the cargo planning and loading operations in an air cargo terminal. The system can assist decision-making process during cargo load planning in the complex operations of air cargo terminal operations. The processes of cargo loading, cargo build-up, security screening, and system monitoring can be further visualised. Scenarios are designed to support and demonstrate the daily operations of the air cargo terminal, including dangerous goods, pets and animals, and some special cargos.

Keywords: air cargo load planning, optimisation, virtual reality, weight and balance, unit load device

Procedia PDF Downloads 338
4214 The Effectiveness of the Recovering from Child Abuse Programme (RCAP) for the Treatment of CPTSD: A Pilot Study

Authors: Siobhan Hegarty, Michael Bloomfield, Kim Entholt, Dorothy Williams, Helen Kennerley

Abstract:

Complex Post-Traumatic Stress Disorder (CPTSD) confers greater risk of poor outcomes than does Post-Traumatic Stress Disorder (PTSD). Despite this, the current treatment guidelines for CPTSD aim to reduce only the ‘core’ symptoms of re-experiencing, hyper-vigilance and avoidance, while not addressing the Disturbances of Self Organisation (DSO) symptoms that distinguish this novel diagnosis from PTSD. The Recovering from Child Abuse Programme (RCAP) is a group protocol, based on the principles of cognitive behavioural therapy (CBT). Preliminary evidence suggests the program is effective at reducing DSO symptoms. This pilot study is the first to investigate the potential effectiveness of the RCAP for the specific treatment of CPTSD. This study was conducted as a service evaluation in a secondary care, traumatic stress service. Treatment was delivered once a week, in two-hour sessions, to ten existing female CPTSD patients of the service, who had experienced sexual abuse in childhood. The programme was administered by two therapists and two additional facilitators, following the RCAP protocol manual. Symptom severity was measured before the administration of therapy and was tracked across a range of measures (International Trauma Questionnaire; Patient Health Questionnaire; Community Assessment of Psychic Experience; Work and Social Adjustment Scale) at five time points, over the course of treatment. Qualitative appraisal of the programme was gathered via weekly feedback forms and from audio-taped recordings of verbal feedback given during group sessions. Preliminary results suggest the programme causes a slight reduction in CPTSD and depressive symptom severity and preliminary qualitative analysis suggests that the RCAP is both helpful and acceptable to group members. Final results and conclusions will follow completed thematic analysis of results.

Keywords: Child sexual abuse, Cognitive behavioural therapy, Complex post-traumatic stress disorder, Recovering from child abuse programme

Procedia PDF Downloads 128
4213 Determination of Optical Constants of Semiconductor Thin Films by Ellipsometry

Authors: Aïssa Manallah, Mohamed Bouafia

Abstract:

Ellipsometry is an optical method based on the study of the behavior of polarized light. The light reflected on a surface induces a change in the polarization state which depends on the characteristics of the material (complex refractive index and thickness of the different layers constituting the device). The purpose of this work is to determine the optical properties of semiconductor thin films by ellipsometry. This paper describes the experimental aspects concerning the semiconductor samples, the SE400 ellipsometer principle, and the results obtained by direct measurements of ellipsometric parameters and modelling using appropriate software.

Keywords: ellipsometry, optical constants, semiconductors, thin films

Procedia PDF Downloads 296
4212 Rheometer Enabled Study of Tissue/biomaterial Frequency-Dependent Properties

Authors: Polina Prokopovich

Abstract:

Despite the well-established dependence of cartilage mechanical properties on the frequency of the applied load, most research in the field is carried out in either load-free or constant load conditions because of the complexity of the equipment required for the determination of time-dependent properties. These simpler analyses provide a limited representation of cartilage properties thus greatly reducing the impact of the information gathered hindering the understanding of the mechanisms involved in this tissue replacement, development and pathology. More complex techniques could represent better investigative methods, but their uptake in cartilage research is limited by the highly specialised training required and cost of the equipment. There is, therefore, a clear need for alternative experimental approaches to cartilage testing to be deployed in research and clinical settings using more user-friendly and financial accessible devices. Frequency dependent material properties can be determined through rheometry that is an easy to use requiring a relatively inexpensive device; we present how a commercial rheometer can be adapted to determine the viscoelastic properties of articular cartilage. Frequency-sweep tests were run at various applied normal loads on immature, mature and trypsinased (as model of osteoarthritis) cartilage samples to determine the dynamic shear moduli (G*, G′ G″) of the tissues. Moduli increased with increasing frequency and applied load; mature cartilage had generally the highest moduli and GAG depleted samples the lowest. Hydraulic permeability (KH) was estimated from the rheological data and decreased with applied load; GAG depleted cartilage exhibited higher hydraulic permeability than either immature or mature tissues. The rheometer-based methodology developed was validated by the close comparison of the rheometer-obtained cartilage characteristics (G*, G′, G″, KH) with results obtained with more complex testing techniques available in literature. Rheometry is relatively simpler and does not require highly capital intensive machinery and staff training is more accessible; thus the use of a rheometer would represent a cost-effective approach for the determination of frequency-dependent properties of cartilage for more comprehensive and impactful results for both healthcare professional and R&D.

Keywords: tissue, rheometer, biomaterial, cartilage

Procedia PDF Downloads 64
4211 A Novel Parametric Chaos-Based Switching System PCSS for Image Encryption

Authors: Mohamed Salah Azzaz, Camel Tanougast, Tarek Hadjem

Abstract:

In this paper, a new low-cost image encryption technique is proposed and analyzed. The developed chaos-based key generator provides complex behavior and can change it automatically via a random-like switching rule. The designed encryption scheme is called PCSS (Parametric Chaos-based Switching System). The performances of this technique were evaluated in terms of data security and privacy. Simulation results have shown the effectiveness of this technique, and it can thereafter, ready for a hardware implementation.

Keywords: chaos, encryption, security, image

Procedia PDF Downloads 462
4210 Processing Design of Miniature Casting Incorporating Stereolithography Technologies

Authors: Pei-Hsing Huang, Wei-Ju Huang

Abstract:

Investment casting is commonly used in the production of metallic components with complex shapes, due to its high dimensional precision, good surface finish, and low cost. However, the process is cumbersome, and the period between trial casting and final production can be very long, thereby limiting business opportunities and competitiveness. In this study, we replaced conventional wax injection with stereolithography (SLA) 3D printing to speed up the trial process and reduce costs. We also used silicone molds to further reduce costs to avoid the high costs imposed by photosensitive resin.

Keywords: investment casting, stereolithography, wax molding, 3D printing

Procedia PDF Downloads 392
4209 Streamwise Vorticity in the Wake of a Sliding Bubble

Authors: R. O’Reilly Meehan, D. B. Murray

Abstract:

In many practical situations, bubbles are dispersed in a liquid phase. Understanding these complex bubbly flows is therefore a key issue for applications such as shell and tube heat exchangers, mineral flotation and oxidation in water treatment. Although a large body of work exists for bubbles rising in an unbounded medium, that of bubbles rising in constricted geometries has received less attention. The particular case of a bubble sliding underneath an inclined surface is common to two-phase flow systems. The current study intends to expand this knowledge by performing experiments to quantify the streamwise flow structures associated with a single sliding air bubble under an inclined surface in quiescent water. This is achieved by means of two-dimensional, two-component particle image velocimetry (PIV), performed with a continuous wave laser and high-speed camera. PIV vorticity fields obtained in a plane perpendicular to the sliding surface show that there is significant bulk fluid motion away from the surface. The associated momentum of the bubble means that this wake motion persists for a significant time before viscous dissipation. The magnitude and direction of the flow structures in the streamwise measurement plane are found to depend on the point on its path through which the bubble enters the plane. This entry point, represented by a phase angle, affects the nature and strength of the vortical structures. This study reconstructs the vorticity field in the wake of the bubble, converting the field at different instances in time to slices of a large-scale wake structure. This is, in essence, Taylor’s ”frozen turbulence” hypothesis. Applying this to the vorticity fields provides a pseudo three-dimensional representation from 2-D data, allowing for a more intuitive understanding of the bubble wake. This study provides insights into the complex dynamics of a situation common to many engineering applications, particularly shell and tube heat exchangers in the nucleate boiling regime.

Keywords: bubbly flow, particle image velocimetry, two-phase flow, wake structures

Procedia PDF Downloads 365
4208 IT System in the Food Supply Chain Safety, Application in SMEs Sector

Authors: Mohsen Shirani, Micaela Demichela

Abstract:

Food supply chain is one of the most complex supply chain networks due to its perishable nature and customer oriented products, and food safety is the major concern for this industry. IT system could help to minimize the production and consumption of unsafe food by controlling and monitoring the entire system. However, there have been many issues in adoption of IT system in this industry specifically within SMEs sector. With this regard, this study presents a novel approach to use IT and tractability systems in the food supply chain, using application of RFID and central database.

Keywords: food supply chain, IT system, safety, SME

Procedia PDF Downloads 464
4207 Synthesis and Characterization of Mixed ligand complexes of Bipyridyl and Glycine with Different Counter Anions as Functional Antioxidant Enzyme Mimics

Authors: Mohamed M. Ibrahim, Gaber A. M. Mersal, Salih Al-Juaid, Samir A. El-Shazly

Abstract:

A series of mixed ligand complexes, viz., [Cu(BPy)(Gly)X]Y {X = Cl (1), Y = 0; X = 0, Y = ClO4- (2); X = H2O, Y = NO3- (3); X = H2O, Y = CH3COO- (4); and [Cu(BPy)(Gly)-(H2O)]2(SO4) (5) have been synthesized. Their structures and properties were characterized by elemental analysis, thermal analaysis, IR, UV–vis, and ESR spectroscopy, as well as electrochemical measurements including cyclic voltammetry, electrical molar conductivity, and magnetic moment measurements. Complexes 1 and 2 formed slightly distorted square-pyramidal coordination geometries of CuN3OCl and CuN3O2, respectively in which the N,O-donor glycine and N,N-donor bipyridyl bind at the basal plane with chloride ion or water as the axial ligand. Complex 3 shows square planar CuN3O coordination geometry, which exhibits chemically significant hydrogen bonding interactions besides showing coordination polymer formation. The superoxide dismutase and catalase-like activities of all complexes were tested and were found to be promising candidates as durable electron-transfer catalyst being close to the efficiency of the mimicking enzymes displaying either catalase or tyrosinase activity to serve for complete reactive oxygen species (ROS) detoxification, both with respect to superoxide radicals and related peroxides. The DNA binding interaction with super coiled pGEM-T plasmid DNA was investigated by using spectral (absorption and emission) titration and electrochemical techniques. The results revealed that DNA intercalate with complexes 1 and 2 through the groove binding mode. The calculated intrinsic binding constant (Kb) of 1 and 2 were 4.71 and 2.429 × 105 M−1, respectively. Gel electrophoresis study reveals the fact that both complexes cleave super coiled pGEM-T plasmid DNA to nicked and linear forms in the absence of any additives. On the other hand, the interaction of both complexes with DNA, the quasi-reversible CuII/CuI redox couple slightly improves its reversibility with considerable decrease in current intensity. All the experimental results indicate that the bipyridyl mixed copper(II) complex (1) intercalate more effectively into the DNA base pairs.

Keywords: enzyme mimics, mixed ligand complexes, X-ray structures, antioxidant, DNA-binding, DNA cleavage

Procedia PDF Downloads 535
4206 Berry Phase and Quantum Skyrmions: A Loop Tour in Physics

Authors: Sinuhé Perea Puente

Abstract:

In several physics systems the whole can be obtained as an exact copy of each of its parts, which facilitates the study of a complex system by looking carefully at its elements, separately. Reducionism offers simplified models which makes the problems easier, but “there’s plenty of room...at the mesoscopic scale”. Here we present a tour for two of its representants: Berry phase and skyrmions, studying some of its basic definitions and properties, and two cases in which both arise together, to finish constraining the scale for our mesoscopic system in the quest of quantum skyrmions, discovering which properties are conserved and which others may be destroyed.

Keywords: condensed mattter, quantum physics, skyrmions, topological defects

Procedia PDF Downloads 126