Search results for: synthetic gene network
807 Expert System: Debugging Using MD5 Process Firewall
Authors: C. U. Om Kumar, S. Kishore, A. Geetha
Abstract:
An Operating system (OS) is software that manages computer hardware and software resources by providing services to computer programs. One of the important user expectations of the operating system is to provide the practice of defending information from unauthorized access, disclosure, modification, inspection, recording or destruction. Operating system is always vulnerable to the attacks of malwares such as computer virus, worm, Trojan horse, backdoors, ransomware, spyware, adware, scareware and more. And so the anti-virus software were created for ensuring security against the prominent computer viruses by applying a dictionary based approach. The anti-virus programs are not always guaranteed to provide security against the new viruses proliferating every day. To clarify this issue and to secure the computer system, our proposed expert system concentrates on authorizing the processes as wanted and unwanted by the administrator for execution. The Expert system maintains a database which consists of hash code of the processes which are to be allowed. These hash codes are generated using MD5 message-digest algorithm which is a widely used cryptographic hash function. The administrator approves the wanted processes that are to be executed in the client in a Local Area Network by implementing Client-Server architecture and only the processes that match with the processes in the database table will be executed by which many malicious processes are restricted from infecting the operating system. The add-on advantage of this proposed Expert system is that it limits CPU usage and minimizes resource utilization. Thus data and information security is ensured by our system along with increased performance of the operating system.Keywords: virus, worm, Trojan horse, back doors, Ransomware, Spyware, Adware, Scareware, sticky software, process table, MD5, CPU usage and resource utilization
Procedia PDF Downloads 427806 Multi-Particle Finite Element Modelling Simulation Based on Cohesive Zone Method of Cold Compaction Behavior of Laminar Al and NaCl Composite Powders
Authors: Yanbing Feng, Deqing Mei, Yancheng Wang, Zichen Chen
Abstract:
With the advantage of low volume density, high specific surface area, light weight and good permeability, porous aluminum material has the potential to be used in automotive, railway, chemistry and construction industries, etc. A layered powder sintering and dissolution method were developed to fabricate the porous surface Al structure with high efficiency. However, the densification mechanism during the cold compaction of laminar composite powders is still unclear. In this study, multi particle finite element modelling (MPFEM) based on the cohesive zone method (CZM) is used to simulate the cold compaction behavior of laminar Al and NaCl composite powders. To obtain its densification mechanism, the macro and micro properties of final compacts are characterized and analyzed. The robustness and accuracy of the numerical model is firstly verified by experimental results and data fitting. The results indicate that the CZM-based multi particle FEM is an effective way to simulate the compaction of the laminar powders and the fracture process of the NaCl powders. In the compaction of the laminar powders, the void is mainly filled by the particle rearrangement, plastic deformation of Al powders and brittle fracture of NaCl powders. Large stress is mainly concentrated within the NaCl powers and the contact force network is formed. The Al powder near the NaCl powder or the mold has larger stress distribution on its contact surface. Therefore, the densification process of cold compaction of laminar Al and NaCl composite powders is successfully analyzed by the CZM-based multi particle FEM.Keywords: cold compaction, cohesive zone, multi-particle FEM, numerical modeling, powder forming
Procedia PDF Downloads 152805 Social Networks And Social Complexity: The Southern Italian Drive For Trade Exchange During The Late Bronze Age
Authors: Sara Fioretti
Abstract:
During the Middle Bronze Age, southern Italy underwent a reorganisation of social structures where local cultures, such as the sub-Apennine and Nuragic, flourished and participated in maritime trade. This paper explores the socio-economic relationships, in both cross-cultural and potentially inter-regional settings, present within the archaeological repertoire of the southern Italian Late Bronze Age (LBA 1600 -1050 BCE). The emergence of economic relations within the connectivity of the regional settlements is explored through ceramic contexts found in the case studies Punta di Zambrone, Broglio di Trebisacce, and Nuraghe Antigori. This paper discusses the findings of a statistical and theoretical approach from an ongoing study in relation to the Mediterranean’s characterisation as a period dominated by Mycenaean influence. This study engages with a theoretical bricolage of Social Networks Entanglement, and Assertive Objects Theory to address the selective and assertive dynamics evident in the cross-cultural trade exchanges as well as consider inter-regional dynamics. Through this intersection of theory and statistical analysis, the case studies establish a small percentage of pottery as imported, whilst assertive productions have a relatively higher quantity. Overall, the majority still adheres to regional Italian traditions. Therefore, we can dissect the rhizomatic relationships cultivated by the Italian coasts and Mycenaeans and their roles within their networks through the intersection of theoretical and statistical analysis. This research offers a new perspective on the complex nature of the Late Bronze Age relational structures.Keywords: late bronze age, mediterranean archaeology, exchanges and trade, frequency distribution of ceramic assemblages, social network theory, rhizomatic exchanges
Procedia PDF Downloads 47804 Job in Modern Arabic Poetry: A Semantic and Comparative Approach to Two Poems Referring to the Poet Al-Sayyab
Authors: Jeries Khoury
Abstract:
The use of legendary, folkloric and religious symbols is one of the most important phenomena in modern Arabic poetry. Interestingly enough, most of the modern Arabic poetry’s pioneers were so fascinated by the biblical symbols and they managed to use many modern techniques to make these symbols adequate for their personal life from one side and fit to their Islamic beliefs from the other. One of the most famous poets to do so was al-Sayya:b. The way he employed one of these symbols ‘job’, the new features he adds to this character and the link between this character and his personal life will be discussed in this study. Besides, the study will examine the influence of al-Sayya:b on another modern poet Saadi Yusuf, who, following al-Sayya:b, used the character of Job in a special way, by mixing its features with al-Sayya:b’s personal features and in this way creating a new mixed character. A semantic, cultural and comparative analysis of the poems written by al-Sayya:b himself and the other poets who evoked the mixed image of al-Sayya:b-Job, can reveal the changes Arab poets made to the original biblical figure of Job to bring it closer to Islamic culture. The paper will make an intensive use of intertextuality idioms in order to shed light on the network of relations between three kinds of texts (indeed three ‘palimpsests’: 1- biblical- the primary text; 2- poetic- al-Syya:b’s secondary version; 3- re-poetic- Sa’di Yusuf’s tertiary version). The bottom line in this paper is that that al-Sayya:b was directly influenced by the dramatic biblical story of Job more than the brief Quranic version of the story. In fact, the ‘new’ character of Job designed by al-Sayya:b himself differs from the original one in many aspects that we can safely say it is the Sayyabian-Job that cannot be found in the poems of any other poets, unless they are evoking the own tragedy of al-Sayya:b himself, like what Saadi Yusuf did.Keywords: Arabic poetry, intertextuality, job, meter, modernism, symbolism
Procedia PDF Downloads 200803 Characterization of the Lytic Bacteriophage VbɸAB-1 against Drug Resistant Acinetobacter baumannii Isolated from Hospitalized Pressure Ulcers Patients
Authors: M. Doudi, M. H. Pazandeh, L. Rahimzadeh Torabi
Abstract:
Bedsores are pressure ulcers that occur on the skin or tissue due to being immobile and lying in bed for extended periods. Bedsores have the potential to progress into open ulcers, increasing the possibility of variety of bacterial infection. Acinetobacter baumannii, a pathogen of considerable clinical importance, exhibited a significant correlation with Bedsores (pressure ulcers) infections, thereby manifesting a wide spectrum of antibiotic resistance. The emergence of drug resistance has led researchers to focus on alternative methods, particularly phage therapy, for tackling bacterial infections. Phage therapy has emerged as a novel therapeutic approach to regulate the activity of these agents. The management of bacterial infections greatly benefits from the clinical utilization of bacteriophages as a valuable antimicrobial intervention. The primary objective of this investigation consisted of isolating and discerning potent bacteriophage capable of targeting multi drug-resistant (MDR) and extensively drug-resistant (XDR) bacteria obtained from pressure ulcers. In present study, analyzed and isolated A. baumannii strains obtained from a cohort of patients suffering from pressure ulcers at Taleghani Hospital in Ahvaz, Iran. An approach that included biochemical and molecular identification techniques was used to determine the taxonomic classification of bacterial isolates at the genus and species levels. The molecular identification process was facilitated by using the 16S rRNA gene in combination with universal primers 27 F, and 1492 R. Bacteriophage was obtained through the isolation process conducted on treatment plant sewage located in Isfahan, Iran. The main goal of this study was to evaluate different characteristics of phage, such as their appearance, range of hosts they can infect, how quickly they can enter a host, their stability at varying temperatures and pH levels, their effectiveness in killing bacteria, the growth pattern of a single phage stage, mapping of enzymatic digestion, and identification of proteomics patterns. The findings demonstrated that an examination was conducted on a sample of 50 specimens, wherein 15 instances of A. baumannii were identified. These microorganisms are the predominant Gram-negative agents known to cause wound infections in individuals suffering from bedsores. The study's findings indicated a high prevalence of antibiotic resistance in the strains isolated from pressure ulcers, excluding the clinical strains that exhibited responsiveness to colistin.According to the findings obtained from assessments of host range and morphological characteristics of bacteriophage VbɸAB-1, it can be concluded that this phage possesses specificity towards A. Baumannii BAH_Glau1001 was classified as a member of the Plasmaviridae family. The bacteriophage mentioned earlier showed the strongest antibacterial effect at a temperature of 18 °C and a pH of 6.5. Through the utilization of sodium dodecyl sulfate-polyacrylamide gel electrophoresis (SDS-PAGE) analysis on protein fragments, it was established that the bacteriophage VbɸAB-1 exhibited a size range between 50 and 75 kilodaltons (KDa). The numerous research findings on the effectiveness of phages and the safety studies conducted suggest that the phages studied in this research can be considered as a practical solution and recommended approach for controlling and treating stubborn pathogens in burn wounds among hospitalized patients.Keywords: acinetobacter baumannii, extremely drug- resistant, phage therapy, surgery wound
Procedia PDF Downloads 93802 Networking Approach for Historic Urban Landscape: Case Study of the Porcelain Capital of China
Abstract:
This article presents a “networking approach” as an alternative to the “layering model” in the issue of the historic urban landscape [HUL], based on research conducted in the historic city of Jingdezhen, the center of the porcelain industry in China. This study points out that the existing HUL concept, which can be traced back to the fundamental conceptual divisions set forth by western science, tends to analyze the various elements of urban heritage (composed of hybrid natural-cultural elements) by layers and ignore the nuanced connections and interweaving structure of various elements. Instead, the networking analysis approach can respond to the challenges of complex heritage networks and to the difficulties that are often faced when modern schemes of looking and thinking of landscape in the Eurocentric heritage model encounters local knowledge of Chinese settlement. The fieldwork in this paper examines the local language regarding place names and everyday uses of urban spaces, thereby highlighting heritage systems grounded in local life and indigenous knowledge. In the context of Chinese “Fengshui”, this paper demonstrates the local knowledge of nature and local intelligence of settlement location and design. This paper suggests that industrial elements (kilns, molding rooms, piers, etc.) and spiritual elements (temples for ceramic saints or water gods) are located in their intimate natural networks. Furthermore, the functional, spiritual, and natural elements are perceived as a whole and evolve as an interactive system. This paper proposes a local and cognitive approach in heritage, which was initially developed in European Landscape Convention and historic landscape characterization projects, and yet seeks a more tentative and nuanced model based on urban ethnography in a Chinese city.Keywords: Chinese city, historic urban landscape, heritage conservation, network
Procedia PDF Downloads 140801 Cross-Dipole Right-Hand Circularly Polarized UHF/VHF Yagi-Uda Antenna for Satellite Applications
Authors: Shativel S., Chandana B. R., Kavya B. C., Obli B. Vikram, Suganthi J., Nagendra Rao G.
Abstract:
Satellite communication plays a pivotal role in modern global communication networks, serving as a vital link between terrestrial infrastructure and remote regions. The demand for reliable satellite reception systems, especially in UHF (Ultra High Frequency) and VHF (Very High Frequency) bands, has grown significantly over the years. This research paper presents the design and optimization of a high-gain, dual-band crossed Yagi-Uda antenna in CST Studio Suite, specifically tailored for satellite reception. The proposed antenna system incorporates a circularly polarized (Right-Hand Circular Polarization - RHCP) design to reduce Faraday loss. Our aim was to use fewer elements and achieve gain, so the antenna is constructed using 6x2 elements arranged in cross dipole and supported with a boom. We have achieved 10.67dBi at 146MHz and 9.28dBi at 437.5MHz.The process includes parameter optimization and fine-tuning of the Yagi-Uda array’s elements, such as the length and spacing of directors and reflectors, to achieve high gain and desirable radiation patterns. Furthermore, the optimization process considers the requirements for UHF and VHF frequency bands, ensuring broad frequency coverage for satellite reception. The results of this research are anticipated to significantly contribute to the advancement of satellite reception systems, enhancing their capabilities to reliably connect remote and underserved areas to the global communication network. Through innovative antenna design and simulation techniques, this study seeks to provide a foundation for the development of next-generation satellite communication infrastructure.Keywords: Yagi-Uda antenna, RHCP, gain, UHF antenna, VHF antenna, CST, radiation pattern.
Procedia PDF Downloads 61800 Machine Learning Techniques for COVID-19 Detection: A Comparative Analysis
Authors: Abeer A. Aljohani
Abstract:
COVID-19 virus spread has been one of the extreme pandemics across the globe. It is also referred to as coronavirus, which is a contagious disease that continuously mutates into numerous variants. Currently, the B.1.1.529 variant labeled as omicron is detected in South Africa. The huge spread of COVID-19 disease has affected several lives and has surged exceptional pressure on the healthcare systems worldwide. Also, everyday life and the global economy have been at stake. This research aims to predict COVID-19 disease in its initial stage to reduce the death count. Machine learning (ML) is nowadays used in almost every area. Numerous COVID-19 cases have produced a huge burden on the hospitals as well as health workers. To reduce this burden, this paper predicts COVID-19 disease is based on the symptoms and medical history of the patient. This research presents a unique architecture for COVID-19 detection using ML techniques integrated with feature dimensionality reduction. This paper uses a standard UCI dataset for predicting COVID-19 disease. This dataset comprises symptoms of 5434 patients. This paper also compares several supervised ML techniques to the presented architecture. The architecture has also utilized 10-fold cross validation process for generalization and the principal component analysis (PCA) technique for feature reduction. Standard parameters are used to evaluate the proposed architecture including F1-Score, precision, accuracy, recall, receiver operating characteristic (ROC), and area under curve (AUC). The results depict that decision tree, random forest, and neural networks outperform all other state-of-the-art ML techniques. This achieved result can help effectively in identifying COVID-19 infection cases.Keywords: supervised machine learning, COVID-19 prediction, healthcare analytics, random forest, neural network
Procedia PDF Downloads 92799 A Review of Blog Assisted Language Learning Research: Based on Bibliometric Analysis
Authors: Bo Ning Lyu
Abstract:
Blog assisted language learning (BALL) has been trialed by educators in language teaching with the development of Web 2.0 technology. Understanding the development trend of related research helps grasp the whole picture of the use of blog in language education. This paper reviews current research related to blogs enhanced language learning based on bibliometric analysis, aiming at (1) identifying the most frequently used keywords and their co-occurrence, (2) clustering research topics based on co-citation analysis, (3) finding the most frequently cited studies and authors and (4) constructing the co-authorship network. 330 articles were searched out in Web of Science, 225 peer-viewed journal papers were finally collected according to selection criteria. Bibexcel and VOSviewer were used to visualize the results. Studies reviewed were published between 2005 to 2016, most in the year of 2014 and 2015 (35 papers respectively). The top 10 most frequently appeared keywords are learning, language, blog, teaching, writing, social, web 2.0, technology, English, communication. 8 research themes could be clustered by co-citation analysis: blogging for collaborative learning, blogging for writing skills, blogging in higher education, feedback via blogs, blogging for self-regulated learning, implementation of using blogs in classroom, comparative studies and audio/video blogs. Early studies focused on the introduction of the classroom implementation while recent studies moved to the audio/video blogs from their traditional usage. By reviewing the research related to BALL quantitatively and objectively, this paper reveals the evolution and development trends as well as identifies influential research, helping researchers and educators quickly grasp this field overall and conducting further studies.Keywords: blog, bibliometric analysis, language learning, literature review
Procedia PDF Downloads 211798 Multiple-Channel Piezoelectric Actuated Tunable Optical Filter for WDM Application
Authors: Hailu Dessalegn, T. Srinivas
Abstract:
We propose new multiple-channel piezoelectric (PZT) actuated tunable optical filter based on racetrack multi-ring resonators for wavelength de-multiplexing network applications. We design tunable eight-channel wavelength de-multiplexer consisting of eight cascaded PZT actuated tunable multi-ring resonator filter with a channel spacing of 1.6 nm. The filter for each channel is basically structured on a suspended beam, sandwiched with piezoelectric material and built in integrated ring resonators which are placed on the middle of the beam to gain uniform stress and linearly varying longitudinal strain. A reference single mode serially coupled multi stage racetrack ring resonator with the same radii and coupling length is designed with a line width of 0.8974 nm with a flat top pass band at 1dB of 0.5205 nm and free spectral range of about 14.9 nm. In each channel, a small change in the perimeter of the rings is introduced to establish the shift in resonance wavelength as per the defined channel spacing. As a result, when a DC voltage is applied, the beams will elongate, which involves mechanical deformation of the ring resonators that induces a stress and a strain, which brings a change in refractive index and perimeter of the rings leading to change in the output spectrum shift providing the tunability of central wavelength in each channel. Simultaneous wave length shift as high as 45.54 pm/V has been achieved with negligible tunability variation in the eight channel tunable optical filter proportional to the DC voltage applied in the structure, and it is capable of tuning up to 3.45 nm in each channel with a maximum loss difference of 0.22 dB in the tuning range and out of band rejection ratio of 35 dB, with a low channel crosstalk ≤ 30 dB.Keywords: optical MEMS, piezoelectric (PZT) actuation, tunable optical filter, wavelength de-multiplexer
Procedia PDF Downloads 437797 Experiences of Homophobia, Machismo and Misogyny in Tourist Destinations: A Netnography in a Facebook Community of LGBT Backpackers
Authors: Renan De Caldas Honorato, Ana Augusta Ferreira De Freitas
Abstract:
Homosexuality is still criminalized in a large number of countries. In some of them, being gay or lesbian can even be punished by death. Added to this context, the experiences of social discrimination faced by the LGBT population, including homophobia, machismo and misogyny, cause numerous restrictions throughout their lives. The possibility of confronting these challenges in moments that should be pleasant, such as on a trip or on vacation, is unpleasant, to say the least. In the current scenario of intensifying the use of Social network sites (SNSs) to search for information, including in the tourist area, this work aims to analyze the sharing of tourist experiences with situations of confrontation and perceptions of homophobia, machismo and misogyny, and restrictions suffered in tourist destinations. The fieldwork is a community of LGBT backpackers based on Facebook. Netnography was the core method adopted. A qualitative approach was conducted and 463 publications posted from January to December 2020 were assessed through the computer-mediated discourse analysis (CMDA). The results suggest that these publications exist to identify the potential exposure to these offensive behaviors while traveling. Individuals affirm that the laws, positive or not, in relation to the LGBT public are not the only factors for a place to be defined as safe or not for gay travelers. The social situation of a country and its laws are quite different and this is the main target of these publications. The perception of others about the chosen destination is more important than knowing your rights and the legal status of each country and it also lessens uncertainty, even when they are never totally confident when choosing a travel destination. In certain circumstances, sexual orientation also needs to be protected from the judgment of hosts and residents. The systemic treatment of homophobic behavior and the construction of a more inclusive society are urgent.Keywords: homophobia, hospitality, machismo, misogyny
Procedia PDF Downloads 188796 Building Transparent Supply Chains through Digital Tracing
Authors: Penina Orenstein
Abstract:
In today’s world, particularly with COVID-19 a constant worldwide threat, organizations need greater visibility over their supply chains more than ever before, in order to find areas for improvement and greater efficiency, reduce the chances of disruption and stay competitive. The concept of supply chain mapping is one where every process and route is mapped in detail between each vendor and supplier. The simplest method of mapping involves sourcing publicly available data including news and financial information concerning relationships between suppliers. An additional layer of information would be disclosed by large, direct suppliers about their production and logistics sites. While this method has the advantage of not requiring any input from suppliers, it also doesn’t allow for much transparency beyond the first supplier tier and may generate irrelevant data—noise—that must be filtered out to find the actionable data. The primary goal of this research is to build data maps of supply chains by focusing on a layered approach. Using these maps, the secondary goal is to address the question as to whether the supply chain is re-engineered to make improvements, for example, to lower the carbon footprint. Using a drill-down approach, the end result is a comprehensive map detailing the linkages between tier-one, tier-two, and tier-three suppliers super-imposed on a geographical map. The driving force behind this idea is to be able to trace individual parts to the exact site where they’re manufactured. In this way, companies can ensure sustainability practices from the production of raw materials through the finished goods. The approach allows companies to identify and anticipate vulnerabilities in their supply chain. It unlocks predictive analytics capabilities and enables them to act proactively. The research is particularly compelling because it unites network science theory with empirical data and presents the results in a visual, intuitive manner.Keywords: data mining, supply chain, empirical research, data mapping
Procedia PDF Downloads 175795 Project Production Control (PPC) Implementation for an Offshore Facilities Construction Project
Authors: Muhammad Hakim Bin Mat Tasir, Erwan Shahfizad Hasidan, Hamidah Makmor Bakry, M. Hafiz B. Izhar
Abstract:
Every key performance indicator used to monitor a project’s construction progress emphasizes trade productivity or specific commodity run-down curves. Examples include the productivity of welding by the number of joints completed per day, quantity of NDT (Non-Destructive Tests) inspection per day, etc. This perspective is based on progress and productivity; however, it does not enable a system perspective of how we produce. This paper uses a project production system perspective by which projects are a collection of production systems comprising the interconnected network of processes and operations that represent all the work activities to execute a project from start to finish. Furthermore, it also uses the 5 Levels of production system optimization as a frame. The goal of the paper is to describe the application of Project Production Control (PPC) to control and improve the performance of several production processes associated with the fabrication and assembly of a Central Processing Platform (CPP) Jacket, part of an offshore mega project. More specifically, the fabrication and assembly of buoyancy tanks as they were identified as part of the critical path and required the highest demand for capacity. In total, seven buoyancy tanks were built, with a total estimated weight of 2,200 metric tons. These huge buoyancy tanks were designed to be reversed launching and self-upending of the jacket, easily retractable, and reusable for the next project, ensuring sustainability. Results showed that an effective application of PPC not only positively impacted construction progress and productivity but also exposed sources of detrimental variability as the focus of continuous improvement practices. This approach augmented conventional project management practices, and the results had a high impact on construction scheduling, planning, and control.Keywords: offshore, construction, project management, sustainability
Procedia PDF Downloads 59794 Analysing the Moderating Effect of Customer Loyalty on Long Run Repurchase Intentions
Authors: John Akpesiri Olotewo
Abstract:
One of the controversies in existing marketing literatures is on how to retain existing and new customers to have repurchase intention in the long-run; however, empirical answer to this question is scanty in existing studies. Thus, this study investigates the moderating effect of consumer loyalty on long-run repurchase intentions in telecommunication industry using Lagos State environs. The study adopted field survey research design using questionnaire to elicit responses from 250 respondents who were selected using random and stratified random sampling techniques from the telecommunication industry in Lagos State, Nigeria. The internal consistency of the research instrument was verified using the Cronbach’s alpha, the result of 0.89 implies the acceptability of the internal consistency of the survey instrument. The test of the research hypotheses were analyzed using Pearson Product Method of Correlation (PPMC), simple regression analysis and inferential statistics with the aid of Statistical Package for Social Science version 20.0 (SPSS). The study confirmed that customer satisfaction has a significant relationship with customer loyalty in the telecommunication industry; also Service quality has a significant relationship with customer loyalty to a brand; loyalty programs have a significant relationship with customer loyalty to a network operator in Nigeria and Customer loyalty has a significant effect on the long run repurchase intentions of the customer. The study concluded that one of the determinants of long term profitability of a business entity is the long run repurchase intentions of its customers which hinges on the level of brand loyalty of the customer. Thus, it was recommended that service providers in Nigeria should improve on factors like customer satisfaction, service quality, and loyalty programs in order to increase the loyalty of their customer to their brands thereby increasing their repurchase intentions.Keywords: customer loyalty, long run repurchase intentions, brands, service quality and customer satisfaction
Procedia PDF Downloads 233793 Reimagining the Management of Telco Supply Chain with Blockchain
Authors: Jeaha Yang, Ahmed Khan, Donna L. Rodela, Mohammed A. Qaudeer
Abstract:
Traditional supply chain silos still exist today due to the difficulty of establishing trust between various partners and technological barriers across industries. Companies lose opportunities and revenue and inadvertently make poor business decisions resulting in further challenges. Blockchain technology can bring a new level of transparency through sharing information with a distributed ledger in a decentralized manner that creates a basis of trust for business. Blockchain is a loosely coupled, hub-style communication network in which trading partners can work indirectly with each other for simpler integration, but they work together through the orchestration of their supply chain operations under a coherent process that is developed jointly. A Blockchain increases efficiencies, lowers costs, and improves interoperability to strengthen and automate the supply chain management process while all partners share the risk. Blockchain ledger is built to track inventory lifecycle for supply chain transparency and keeps a journal of inventory movement for real-time reconciliation. State design patterns are used to capture the life cycle (behavior) of inventory management as a state machine for a common, transparent and coherent process which creates an opportunity for trading partners to become more responsive in terms of changes or improvements in process, reconcile discrepancies, and comply with internal governance and external regulations. It enables end-to-end, inter-company visibility at the unit level for more accurate demand planning with better insight into order fulfillment and replenishment.Keywords: supply chain management, inventory trace-ability, perpetual inventory system, inventory lifecycle, blockchain, inventory consignment, supply chain transparency, digital thread, demand planning, hyper ledger fabric
Procedia PDF Downloads 90792 The Impact of PM-Based Regulations on the Concentration and Sources of Fine Organic Carbon in the Los Angeles Basin from 2005 to 2015
Authors: Abdulmalik Altuwayjiri, Milad Pirhadi, Sina Taghvaee, Constantinos Sioutas
Abstract:
A significant portion of PM₂.₅ mass concentration is carbonaceous matter (CM), which majorly exists in the form of organic carbon (OC). Ambient OC originates from a multitude of sources and plays an important role in global climate effects, visibility degradation, and human health. In this study, positive matrix factorization (PMF) was utilized to identify and quantify the long-term contribution of PM₂.₅ sources to total OC mass concentration in central Los Angeles (CELA) and Riverside (i.e., receptor site), using the chemical speciation network (CSN) database between 2005 and 2015, a period during which several state and local regulations on tailpipe emissions were implemented in the area. Our PMF resolved five different factors, including tailpipe emissions, non-tailpipe emissions, biomass burning, secondary organic aerosol (SOA), and local industrial activities for both sampling sites. The contribution of vehicular exhaust emissions to the OC mass concentrations significantly decreased from 3.5 µg/m³ in 2005 to 1.5 µg/m³ in 2015 (by about 58%) at CELA, and from 3.3 µg/m³ in 2005 to 1.2 µg/m³ in 2015 (by nearly 62%) at Riverside. Additionally, SOA contribution to the total OC mass, showing higher levels at the receptor site, increased from 23% in 2005 to 33% and 29% in 2010 and 2015, respectively, in Riverside, whereas the corresponding contribution at the CELA site was 16%, 21% and 19% during the same period. The biomass burning maintained an almost constant relative contribution over the whole period. Moreover, while the adopted regulations and policies were very effective at reducing the contribution of tailpipe emissions, they have led to an overall increase in the fractional contributions of non-tailpipe emissions to total OC in CELA (about 14%, 28%, and 28% in 2005, 2010 and 2015, respectively) and Riverside (22%, 27% and 26% in 2005, 2010 and 2015), underscoring the necessity to develop equally effective mitigation policies targeting non-tailpipe PM emissions.Keywords: PM₂.₅, organic carbon, Los Angeles megacity, PMF, source apportionment, non-tailpipe emissions
Procedia PDF Downloads 198791 Present-Day Transformations and Trends in Rooftop Agriculture and Food Security
Authors: Kiara Lawrence, Nadine Ponnusamy, Clive Greenstone
Abstract:
One of the major challenges facing society today is food security. The risks to food security have increased significantly due to the evolving urban landscape, globalization, and a rising population. The cultivation of food is essential, particularly during times of crisis, such as a recession, and has long been a necessity for urban populations. In contemporary society, many urban residents are confronted with new challenges, including high levels of unemployment, which compel individuals to adopt alternative survival strategies, such as growing their own food. Recently, rooftop agriculture has made significant contributions to urban and national food security and has been utilized as a tool to mitigate the frequent and damaging disasters that many cities encounter. They have the potential to transform unused spaces into green, productive vegetable plots, while also providing urban residents with the opportunity to enjoy the benefits of gardening. Therefore, this study looks to investigate the evolving themes around rooftop agriculture and food security globally. A bibliometric review analysis was carried out on Scopus and Web of Science using the keywords “rooftop agriculture” OR “rooftop farming” OR “rooftop garden” AND “food security” between 2004 and 2024 to ensure a broader scope was covered around the chosen study. Vosviewer software was then utilized to analyze the extracted data to create network visualization maps based on keyword occurrences, co-author analysis, country analysis. There were only 37 relevant documents within the study parameters. Preliminary results indicate that much research focused on urban agriculture, food supply, green roof, sustainability and climate change. By analysing these aspects of rooftop agriculture and food security, the trends can identify gaps in literature and dictate future applications to assist in food security.Keywords: food security, rooftop agriculture, rooftop farming, rooftop garden
Procedia PDF Downloads 17790 Deterministic and Stochastic Modeling of a Micro-Grid Management for Optimal Power Self-Consumption
Authors: D. Calogine, O. Chau, S. Dotti, O. Ramiarinjanahary, P. Rasoavonjy, F. Tovondahiniriko
Abstract:
Mafate is a natural circus in the north-western part of Reunion Island, without an electrical grid and road network. A micro-grid concept is being experimented in this area, composed of a photovoltaic production combined with electrochemical batteries, in order to meet the local population for self-consumption of electricity demands. This work develops a discrete model as well as a stochastic model in order to reach an optimal equilibrium between production and consumptions for a cluster of houses. The management of the energy power leads to a large linearized programming system, where the time interval of interest is 24 hours The experimental data are solar production, storage energy, and the parameters of the different electrical devices and batteries. The unknown variables to evaluate are the consumptions of the various electrical services, the energy drawn from and stored in the batteries, and the inhabitants’ planning wishes. The objective is to fit the solar production to the electrical consumption of the inhabitants, with an optimal use of the energies in the batteries by satisfying as widely as possible the users' planning requirements. In the discrete model, the different parameters and solutions of the linear programming system are deterministic scalars. Whereas in the stochastic approach, the data parameters and the linear programming solutions become random variables, then the distributions of which could be imposed or established by estimation from samples of real observations or from samples of optimal discrete equilibrium solutions.Keywords: photovoltaic production, power consumption, battery storage resources, random variables, stochastic modeling, estimations of probability distributions, mixed integer linear programming, smart micro-grid, self-consumption of electricity.
Procedia PDF Downloads 110789 Functionalized Nano porous Ceramic Membranes for Electrodialysis Treatment of Harsh Wastewater
Authors: Emily Rabe, Stephanie Candelaria, Rachel Malone, Olivia Lenz, Greg Newbloom
Abstract:
Electrodialysis (ED) is a well-developed technology for ion removal in a variety of applications. However, many industries generate harsh wastewater streams that are incompatible with traditional ion exchange membranes. Membrion® has developed novel ceramic-based ion exchange membranes (IEMs) offering several advantages over traditional polymer membranes: high performance in low pH, chemical resistance to oxidizers, and a rigid structure that minimizes swelling. These membranes are synthesized with our patented silane-based sol-gel techniques. The pore size, shape, and network structure are engineered through a molecular self-assembly process where thermodynamic driving forces are used to direct where and how pores form. Either cationic or anionic groups can be added within the membrane nanopore structure to create cation- and anion-exchange membranes. The ceramic IEMs are produced on a roll-to-roll manufacturing line with low-temperature processing. Membrane performance testing is conducted using in-house permselectivity, area-specific resistance, and ED stack testing setups. Ceramic-based IEMs show comparable performance to traditional IEMs and offer some unique advantages. Long exposure to highly acidic solutions has a negligible impact on ED performance. Additionally, we have observed stable performance in the presence of strong oxidizing agents such as hydrogen peroxide. This stability is expected, as the ceramic backbone of these materials is already in a fully oxidized state. This data suggests ceramic membranes, made using sol-gel chemistry, could be an ideal solution for acidic and/or oxidizing wastewater streams from processes such as semiconductor manufacturing and mining.Keywords: ion exchange, membrane, silane chemistry, nanostructure, wastewater
Procedia PDF Downloads 86788 Detect Critical Thinking Skill in Written Text Analysis. The Use of Artificial Intelligence in Text Analysis vs Chat/Gpt
Authors: Lucilla Crosta, Anthony Edwards
Abstract:
Companies and the market place nowadays struggle to find employees with adequate skills in relation to anticipated growth of their businesses. At least half of workers will need to undertake some form of up-skilling process in the next five years in order to remain aligned with the requests of the market . In order to meet these challenges, there is a clear need to explore the potential uses of AI (artificial Intelligence) based tools in assessing transversal skills (critical thinking, communication and soft skills of different types in general) of workers and adult students while empowering them to develop those same skills in a reliable trustworthy way. Companies seek workers with key transversal skills that can make a difference between workers now and in the future. However, critical thinking seems to be the one of the most imprtant skill, bringing unexplored ideas and company growth in business contexts. What employers have been reporting since years now, is that this skill is lacking in the majority of workers and adult students, and this is particularly visible trough their writing. This paper investigates how critical thinking and communication skills are currently developed in Higher Education environments through use of AI tools at postgraduate levels. It analyses the use of a branch of AI namely Machine Learning and Big Data and of Neural Network Analysis. It also examines the potential effect the acquisition of these skills through AI tools and what kind of effects this has on employability This paper will draw information from researchers and studies both at national (Italy & UK) and international level in Higher Education. The issues associated with the development and use of one specific AI tool Edulai, will be examined in details. Finally comparisons will be also made between these tools and the more recent phenomenon of Chat GPT and forthcomings and drawbacks will be analysed.Keywords: critical thinking, artificial intelligence, higher education, soft skills, chat GPT
Procedia PDF Downloads 110787 Application of a Confirmatory Composite Model for Assessing the Extent of Agricultural Digitalization: A Case of Proactive Land Acquisition Strategy (PLAS) Farmers in South Africa
Authors: Mazwane S., Makhura M. N., Ginege A.
Abstract:
Digitalization in South Africa has received considerable attention from policymakers. The support for the development of the digital economy by the South African government has been demonstrated through the enactment of various national policies and strategies. This study sought to develop an index for agricultural digitalization by applying composite confirmatory analysis (CCA). Another aim was to determine the factors that affect the development of digitalization in PLAS farms. Data on the indicators of the three dimensions of digitalization were collected from 300 Proactive Land Acquisition Strategy (PLAS) farms in South Africa using semi-structured questionnaires. Confirmatory composite analysis (CCA) was employed to reduce the items into three digitalization dimensions and ultimately to a digitalization index. Standardized digitalization index scores were extracted and fitted to a linear regression model to determine the factors affecting digitalization development. The results revealed that the model shows practical validity and can be used to measure digitalization development as measures of fit (geodesic distance, standardized root mean square residual, and squared Euclidean distance) were all below their respective 95%quantiles of bootstrap discrepancies (HI95 values). Therefore, digitalization is an emergent variable that can be measured using CCA. The average level of digitalization in PLAS farms was 0.2 and varied significantly across provinces. The factors that significantly influence digitalization development in PLAS land reform farms were age, gender, farm type, network type, and cellular data type. This should enable researchers and policymakers to understand the level of digitalization and patterns of development, as well as correctly attribute digitalization development to the contributing factors.Keywords: agriculture, digitalization, confirmatory composite model, land reform, proactive land acquisition strategy, South Africa
Procedia PDF Downloads 63786 An Evolutionary Perspective on the Role of Extrinsic Noise in Filtering Transcript Variability in Small RNA Regulation in Bacteria
Authors: Rinat Arbel-Goren, Joel Stavans
Abstract:
Cell-to-cell variations in transcript or protein abundance, called noise, may give rise to phenotypic variability between isogenic cells, enhancing the probability of survival under stress conditions. These variations may be introduced by post-transcriptional regulatory processes such as non-coding, small RNAs stoichiometric degradation of target transcripts in bacteria. We study the iron homeostasis network in Escherichia coli, in which the RyhB small RNA regulates the expression of various targets as a model system. Using fluorescence reporter genes to detect protein levels and single-molecule fluorescence in situ hybridization to monitor transcripts levels in individual cells, allows us to compare noise at both transcript and protein levels. The experimental results and computer simulations show that extrinsic noise buffers through a feed-forward loop configuration the increase in variability introduced at the transcript level by iron deprivation, illuminating the important role that extrinsic noise plays during stress. Surprisingly, extrinsic noise also decouples of fluctuations of two different targets, in spite of RyhB being a common upstream factor degrading both. Thus, phenotypic variability increases under stress conditions by the decoupling of target fluctuations in the same cell rather than by increasing the noise of each. We also present preliminary results on the adaptation of cells to prolonged iron deprivation in order to shed light on the evolutionary role of post-transcriptional downregulation by small RNAs.Keywords: cell-to-cell variability, Escherichia coli, noise, single-molecule fluorescence in situ hybridization (smFISH), transcript
Procedia PDF Downloads 164785 Human-Machine Cooperation in Facial Comparison Based on Likelihood Scores
Authors: Lanchi Xie, Zhihui Li, Zhigang Li, Guiqiang Wang, Lei Xu, Yuwen Yan
Abstract:
Image-based facial features can be classified into category recognition features and individual recognition features. Current automated face recognition systems extract a specific feature vector of different dimensions from a facial image according to their pre-trained neural network. However, to improve the efficiency of parameter calculation, an algorithm generally reduces the image details by pooling. The operation will overlook the details concerned much by forensic experts. In our experiment, we adopted a variety of face recognition algorithms based on deep learning, compared a large number of naturally collected face images with the known data of the same person's frontal ID photos. Downscaling and manual handling were performed on the testing images. The results supported that the facial recognition algorithms based on deep learning detected structural and morphological information and rarely focused on specific markers such as stains and moles. Overall performance, distribution of genuine scores and impostor scores, and likelihood ratios were tested to evaluate the accuracy of biometric systems and forensic experts. Experiments showed that the biometric systems were skilled in distinguishing category features, and forensic experts were better at discovering the individual features of human faces. In the proposed approach, a fusion was performed at the score level. At the specified false accept rate, the framework achieved a lower false reject rate. This paper contributes to improving the interpretability of the objective method of facial comparison and provides a novel method for human-machine collaboration in this field.Keywords: likelihood ratio, automated facial recognition, facial comparison, biometrics
Procedia PDF Downloads 130784 Adapting Tools for Text Monitoring and for Scenario Analysis Related to the Field of Social Disasters
Authors: Svetlana Cojocaru, Mircea Petic, Inga Titchiev
Abstract:
Humanity faces more and more often with different social disasters, which in turn can generate new accidents and catastrophes. To mitigate their consequences, it is important to obtain early possible signals about the events which are or can occur and to prepare the corresponding scenarios that could be applied. Our research is focused on solving two problems in this domain: identifying signals related that an accident occurred or may occur and mitigation of some consequences of disasters. To solve the first problem, methods of selecting and processing texts from global network Internet are developed. Information in Romanian is of special interest for us. In order to obtain the mentioned tools, we should follow several steps, divided into preparatory stage and processing stage. Throughout the first stage, we manually collected over 724 news articles and classified them into 10 categories of social disasters. It constitutes more than 150 thousand words. Using this information, a controlled vocabulary of more than 300 keywords was elaborated, that will help in the process of classification and identification of the texts related to the field of social disasters. To solve the second problem, the formalism of Petri net has been used. We deal with the problem of inhabitants’ evacuation in useful time. The analysis methods such as reachability or coverability tree and invariants technique to determine dynamic properties of the modeled systems will be used. To perform a case study of properties of extended evacuation system by adding time, the analysis modules of PIPE such as Generalized Stochastic Petri Nets (GSPN) Analysis, Simulation, State Space Analysis, and Invariant Analysis have been used. These modules helped us to obtain the average number of persons situated in the rooms and the other quantitative properties and characteristics related to its dynamics.Keywords: lexicon of disasters, modelling, Petri nets, text annotation, social disasters
Procedia PDF Downloads 197783 Gnss Aided Photogrammetry for Digital Mapping
Authors: Muhammad Usman Akram
Abstract:
This research work based on GNSS-Aided Photogrammetry for Digital Mapping. It focuses on topographic survey of an area or site which is to be used in future Planning & development (P&D) or can be used for further, examination, exploration, research and inspection. Survey and Mapping in hard-to-access and hazardous areas are very difficult by using traditional techniques and methodologies; as well it is time consuming, labor intensive and has less precision with limited data. In comparison with the advance techniques it is saving with less manpower and provides more precise output with a wide variety of multiple data sets. In this experimentation, Aerial Photogrammetry technique is used where an UAV flies over an area and captures geocoded images and makes a Three-Dimensional Model (3-D Model), UAV operates on a user specified path or area with various parameters; Flight altitude, Ground sampling distance (GSD), Image overlapping, Camera angle etc. For ground controlling, a network of points on the ground would be observed as a Ground Control point (GCP) using Differential Global Positioning System (DGPS) in PPK or RTK mode. Furthermore, that raw data collected by UAV and DGPS will be processed in various Digital image processing programs and Computer Aided Design software. From which as an output we obtain Points Dense Cloud, Digital Elevation Model (DEM) and Ortho-photo. The imagery is converted into geospatial data by digitizing over Ortho-photo, DEM is further converted into Digital Terrain Model (DTM) for contour generation or digital surface. As a result, we get Digital Map of area to be surveyed. In conclusion, we compared processed data with exact measurements taken on site. The error will be accepted if the amount of error is not breached from survey accuracy limits set by concerned institutions.Keywords: photogrammetry, post processing kinematics, real time kinematics, manual data inquiry
Procedia PDF Downloads 32782 Investigation of Different Machine Learning Algorithms in Large-Scale Land Cover Mapping within the Google Earth Engine
Authors: Amin Naboureh, Ainong Li, Jinhu Bian, Guangbin Lei, Hamid Ebrahimy
Abstract:
Large-scale land cover mapping has become a new challenge in land change and remote sensing field because of involving a big volume of data. Moreover, selecting the right classification method, especially when there are different types of landscapes in the study area is quite difficult. This paper is an attempt to compare the performance of different machine learning (ML) algorithms for generating a land cover map of the China-Central Asia–West Asia Corridor that is considered as one of the main parts of the Belt and Road Initiative project (BRI). The cloud-based Google Earth Engine (GEE) platform was used for generating a land cover map for the study area from Landsat-8 images (2017) by applying three frequently used ML algorithms including random forest (RF), support vector machine (SVM), and artificial neural network (ANN). The selected ML algorithms (RF, SVM, and ANN) were trained and tested using reference data obtained from MODIS yearly land cover product and very high-resolution satellite images. The finding of the study illustrated that among three frequently used ML algorithms, RF with 91% overall accuracy had the best result in producing a land cover map for the China-Central Asia–West Asia Corridor whereas ANN showed the worst result with 85% overall accuracy. The great performance of the GEE in applying different ML algorithms and handling huge volume of remotely sensed data in the present study showed that it could also help the researchers to generate reliable long-term land cover change maps. The finding of this research has great importance for decision-makers and BRI’s authorities in strategic land use planning.Keywords: land cover, google earth engine, machine learning, remote sensing
Procedia PDF Downloads 113781 Impact of Charging PHEV at Different Penetration Levels on Power System Network
Authors: M. R. Ahmad, I. Musirin, M. M. Othman, N. A. Rahmat
Abstract:
Plug-in Hybrid-Electric Vehicle (PHEV) has gained immense popularity in recent years. PHEV offers numerous advantages compared to the conventional internal-combustion engine (ICE) vehicle. Millions of PHEVs are estimated to be on the road in the USA by 2020. Uncoordinated PHEV charging is believed to cause severe impacts to the power grid; i.e. feeders, lines and transformers overload and voltage drop. Nevertheless, improper PHEV data model used in such studies may cause the findings of their works is in appropriated. Although smart charging is more attractive to researchers in recent years, its implementation is not yet attainable on the street due to its requirement for physical infrastructure readiness and technology advancement. As the first step, it is finest to study the impact of charging PHEV based on real vehicle travel data from National Household Travel Survey (NHTS) and at present charging rate. Due to the lack of charging station on the street at the moment, charging PHEV at home is the best option and has been considered in this work. This paper proposed a technique that comprehensively presents the impact of charging PHEV on power system networks considering huge numbers of PHEV samples with its traveling data pattern. Vehicles Charging Load Profile (VCLP) is developed and implemented in IEEE 30-bus test system that represents a portion of American Electric Power System (Midwestern US). Normalization technique is used to correspond to real time loads at all buses. Results from the study indicated that charging PHEV using opportunity charging will have significant impacts on power system networks, especially whereas bigger battery capacity (kWh) is used as well as for higher penetration level.Keywords: plug-in hybrid electric vehicle, transportation electrification, impact of charging PHEV, electricity demand profile, load profile
Procedia PDF Downloads 287780 Burkholderia Cepacia ST 767 Causing a Three Years Nosocomial Outbreak in a Hemodialysis Unit
Authors: Gousilin Leandra Rocha Da Silva, Stéfani T. A. Dantas, Bruna F. Rossi, Erika R. Bonsaglia, Ivana G. Castilho, Terue Sadatsune, Ary Fernandes Júnior, Vera l. M. Rall
Abstract:
Kidney failure causes decreased diuresis and accumulation of nitrogenous substances in the body. To increase patient survival, hemodialysis is used as a partial substitute for renal function. However, contamination of the water used in this treatment, causing bacteremia in patients, is a worldwide concern. The Burkholderia cepacia complex (Bcc), a group of bacteria with more than 20 species, is frequently isolated from hemodialysis water samples and comprises opportunistic bacteria, affecting immunosuppressed patients, due to its wide variety of virulence factors, in addition to innate resistance to several antimicrobial agents, contributing to the permanence in the hospital environment and to the pathogenesis in the host. The objective of the present work was to characterize molecularly and phenotypically Bcc isolates collected from the water and dialysate of the Hemodialysis Unit and from the blood of patients at a Public Hospital in Botucatu, São Paulo, Brazil, between 2019 and 2021. We used 33 Bcc isolates, previously obtained from blood cultures from patients with bacteremia undergoing hemodialysis treatment (2019-2021) and 24 isolates obtained from water and dialysate samples in a Hemodialysis Unit (same period). The recA gene was sequenced to identify the specific species among the Bcc group. All isolates were tested for the presence of some genes that encode virulence factors such as cblA, esmR, zmpA and zmpB. Considering the epidemiology of the outbreak, the Bcc isolates were molecularly characterized by Multi Locus Sequence Type (MLST) and by pulsed-field gel electrophoresis (PFGE). The verification and quantification of biofilm in a polystyrene microplate were performed by submitting the isolates to different incubation temperatures (20°C, average water temperature and 35°C, optimal temperature for group growth). The antibiogram was performed with disc diffusion tests on agar, using discs impregnated with cefepime (30µg), ceftazidime (30µg), ciprofloxacin (5µg), gentamicin (10µg), imipenem (10µg), amikacin 30µg), sulfametazol/trimethoprim (23.75/1.25µg) and ampicillin/sulbactam (10/10µg). The presence of ZmpB was identified in all isolates, while ZmpA was observed in 96.5% of the isolates, while none of them presented the cblA and esmR genes. The antibiogram of the 33 human isolates indicated that all were resistant to gentamicin, colistin, ampicillin/sulbactam and imipenem. 16 (48.5%) isolates were resistant to amikacin and lower rates of resistance were observed for meropenem, ceftazidime, cefepime, ciprofloxacin and piperacycline/tazobactam (6.1%). All isolates were sensitive to sulfametazol/trimethoprim, levofloxacin and tigecycline. As for the water isolates, resistance was observed only to gentamicin (34.8%) and imipenem (17.4%). According to PFGE results, all isolates obtained from humans and water belonged to the same pulsotype (1), which was identified by recA sequencing as B. cepacia¸, belonging to sequence type ST-767. By observing a single pulse type over three years, one can observe the persistence of this isolate in the pipeline, contaminating patients undergoing hemodialysis, despite the routine disinfection of water with peracetic acid. This persistence is probably due to the production of biofilm, which protects bacteria from disinfectants and, making this scenario more critical, several isolates proved to be multidrug-resistant (resistance to at least three groups of antimicrobials), turning the patient care even more difficult.Keywords: hemodialysis, burkholderia cepacia, PFGE, MLST, multi drug resistance
Procedia PDF Downloads 99779 Toxic Masculinity as Dictatorship: Gender and Power Struggles in Tomás Eloy Martínez´s Novels
Authors: Mariya Dzhyoyeva
Abstract:
In the present paper, I examine manifestations of toxic masculinity in the novels by Tomás Eloy Martínez, a post-Boom author, journalist, literary critic, and one of the representatives of the Argentine writing diaspora. I focus on the analysis of Martínez´s characters that display hypermasculine traits to define the relationship between toxic masculinity and power, including the power of authorship and violence as they are represented in his novels. The analysis reveals a complex network in which gender, power, and violence are intertwined and influence and modify each other. As the author exposes toxic masculine behaviors that generate violence, he looks to undermine them. Departing from M. Kimmel´s idea of masculinity as homophobia, I examine how Martínez “outs” his characters by incorporating into the narrative some secret, privileged sources that provide alternative accounts of their otherwise hypermasculine lives. These background stories expose their “weaknesses,” both physical and mental, and thereby feminize them in their own eyes. In a similar way, the toxic masculinity of the fictional male author that wields his power by abusing the written word as he abuses the female character in the story is exposed as a complex of insecurities accumulated by the character due to his childhood trauma. The artistic technique that Martínez uses to condemn the authoritarian male behavior is accessing his subjectivity and subverting it through a multiplicity of identities. Martínez takes over the character’s “I” and turns it into a host of pronouns with a constantly shifting point of reference that distorts not only the notions of gender but also the very notion of identity. In doing so, he takes the character´s affirmation of masculinity to the limit where the very idea of it becomes unsustainable. Viewed in the context of Martínez´s own exilic story, the condemnation of toxic masculine power turns into the condemnation of dictatorship and authoritarianism.Keywords: gender, masculinity., toxic masculinity, authoritarian, Argentine literature, Martínez
Procedia PDF Downloads 71778 Comparison of Two Neural Networks To Model Margarine Age And Predict Shelf-Life Using Matlab
Authors: Phakamani Xaba, Robert Huberts, Bilainu Oboirien
Abstract:
The present study was aimed at developing & comparing two neural-network-based predictive models to predict shelf-life/product age of South African margarine using free fatty acid (FFA), water droplet size (D3.3), water droplet distribution (e-sigma), moisture content, peroxide value (PV), anisidine valve (AnV) and total oxidation (totox) value as input variables to the model. Brick margarine products which had varying ages ranging from fresh i.e. week 0 to week 47 were sourced. The brick margarine products which had been stored at 10 & 25 °C and were characterized. JMP and MATLAB models to predict shelf-life/ margarine age were developed and their performances were compared. The key performance indicators to evaluate the model performances were correlation coefficient (CC), root mean square error (RMSE), and mean absolute percentage error (MAPE) relative to the actual data. The MATLAB-developed model showed a better performance in all three performance indicators. The correlation coefficient of the MATLAB model was 99.86% versus 99.74% for the JMP model, the RMSE was 0.720 compared to 1.005 and the MAPE was 7.4% compared to 8.571%. The MATLAB model was selected to be the most accurate, and then, the number of hidden neurons/ nodes was optimized to develop a single predictive model. The optimized MATLAB with 10 neurons showed a better performance compared to the models with 1 & 5 hidden neurons. The developed models can be used by margarine manufacturers, food research institutions, researchers etc, to predict shelf-life/ margarine product age, optimize addition of antioxidants, extend shelf-life of products and proactively troubleshoot for problems related to changes which have an impact on shelf-life of margarine without conducting expensive trials.Keywords: margarine shelf-life, predictive modelling, neural networks, oil oxidation
Procedia PDF Downloads 197