Search results for: Markov chain
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2003

Search results for: Markov chain

1133 Fight against Money Laundering with Optical Character Recognition

Authors: Saikiran Subbagari, Avinash Malladhi

Abstract:

Anti Money Laundering (AML) regulations are designed to prevent money laundering and terrorist financing activities worldwide. Financial institutions around the world are legally obligated to identify, assess and mitigate the risks associated with money laundering and report any suspicious transactions to governing authorities. With increasing volumes of data to analyze, financial institutions seek to automate their AML processes. In the rise of financial crimes, optical character recognition (OCR), in combination with machine learning (ML) algorithms, serves as a crucial tool for automating AML processes by extracting the data from documents and identifying suspicious transactions. In this paper, we examine the utilization of OCR for AML and delve into various OCR techniques employed in AML processes. These techniques encompass template-based, feature-based, neural network-based, natural language processing (NLP), hidden markov models (HMMs), conditional random fields (CRFs), binarizations, pattern matching and stroke width transform (SWT). We evaluate each technique, discussing their strengths and constraints. Also, we emphasize on how OCR can improve the accuracy of customer identity verification by comparing the extracted text with the office of foreign assets control (OFAC) watchlist. We will also discuss how OCR helps to overcome language barriers in AML compliance. We also address the implementation challenges that OCR-based AML systems may face and offer recommendations for financial institutions based on the data from previous research studies, which illustrate the effectiveness of OCR-based AML.

Keywords: anti-money laundering, compliance, financial crimes, fraud detection, machine learning, optical character recognition

Procedia PDF Downloads 141
1132 In Silico Analysis of Deleterious nsSNPs (Missense) of Dihydrolipoamide Branched-Chain Transacylase E2 Gene Associated with Maple Syrup Urine Disease Type II

Authors: Zainab S. Ahmed, Mohammed S. Ali, Nadia A. Elshiekh, Sami Adam Ibrahim, Ghada M. El-Tayeb, Ahmed H. Elsadig, Rihab A. Omer, Sofia B. Mohamed

Abstract:

Maple syrup urine (MSUD) is an autosomal recessive disease that causes a deficiency in the enzyme branched-chain alpha-keto acid (BCKA) dehydrogenase. The development of disease has been associated with SNPs in the DBT gene. Despite that, the computational analysis of SNPs in coding and noncoding and their functional impacts on protein level still remains unknown. Hence, in this study, we carried out a comprehensive in silico analysis of missense that was predicted to have a harmful influence on DBT structure and function. In this study, eight different in silico prediction algorithms; SIFT, PROVEAN, MutPred, SNP&GO, PhD-SNP, PANTHER, I-Mutant 2.0 and MUpo were used for screening nsSNPs in DBT including. Additionally, to understand the effect of mutations in the strength of the interactions that bind protein together the ELASPIC servers were used. Finally, the 3D structure of DBT was formed using Mutation3D and Chimera servers respectively. Our result showed that a total of 15 nsSNPs confirmed by 4 software (R301C, R376H, W84R, S268F, W84C, F276C, H452R, R178H, I355T, V191G, M444T, T174A, I200T, R113H, and R178C) were found damaging and can lead to a shift in DBT gene structure. Moreover, we found 7 nsSNPs located on the 2-oxoacid_dh catalytic domain, 5 nsSNPs on the E_3 binding domain and 3 nsSNPs on the Biotin Domain. So these nsSNPs may alter the putative structure of DBT’s domain. Furthermore, we detected all these nsSNPs are on the core residues of the protein and have the ability to change the stability of the protein. Additionally, we found W84R, S268F, and M444T have high significance, and they affected Leucine, Isoleucine, and Valine, which reduces or disrupt the function of BCKD complex, E2-subunit which the DBT gene encodes. In conclusion, based on our extensive in-silico analysis, we report 15 nsSNPs that have possible association with protein deteriorating and disease-causing abilities. These candidate SNPs can aid in future studies on Maple Syrup Urine Disease type II base in the genetic level.

Keywords: DBT gene, ELASPIC, in silico analysis, UCSF chimer

Procedia PDF Downloads 200
1131 Repository Blockchain for Collaborative Blockchain Ecosystem

Authors: Razwan Ahmed Tanvir, Greg Speegle

Abstract:

Collaborative blockchain ecosystems allow diverse groups to cooperate on tasks while providing properties such as decentralization and transaction security. We provide a model that uses a repository blockchain to manage hard forks within a collaborative system such that a single process (assuming that it has knowledge of the requirements of each fork) can access all of the blocks within the system. The repository blockchain replaces the need for Inter Blockchain Communication (IBC) within the ecosystem by navigating the networks. The resulting construction resembles a tree instead of a chain. A proof-of-concept implementation performs a depth-first search on the new structure.

Keywords: hard fork, shared governance, Inter Blockchain Communication, blockchain ecosystem, regular research paper

Procedia PDF Downloads 2
1130 Complex Dynamics in a Morphologically Heterogeneous Biological Medium

Authors: Turky Al-Qahtani, Roustem Miftahof

Abstract:

Introduction: Under common assumptions of excitabi-lity, morphological (cellular) homogeneity, and spatial structural anomalies added as required, it has been shown that biological systems are able to display travelling wave dynamics. Being not self-sustainable, existence depends on the electrophysiological state of transmembrane ion channels and it requires an extrinsic/intrinsic periodic source. However, organs in the body are highly multicellular, heterogeneous, and their functionality is the outcome of electro-mechanical conjugation, rather than excitability only. Thus, peristalsis in the gut relies on spatiotemporal myoelectrical pattern formations between the mechanical, represented by smooth muscle cells (SM), and the control, comprised of a chain of primary sensory and motor neurones, components. Synaptically linked through the afferent and efferent pathways, they form a functional unit (FU) of the gut. Aims: These are: i) to study numerically the complex dynamics, and ii) to investigate the possibility of self-sustained myoelectrical activity in the FU. Methods: The FU recreates the following sequence of physiological events: deformation of mechanoreceptors of located in SM; generation and propagation of electrical waves of depolarisation - spikes - along the axon to the soma of the primary neurone; discharge of the primary neurone and spike propagation towards the motor neurone; burst of the motor neurone and transduction of spikes to SM, subsequently producing forces of contraction. These are governed by a system of nonlinear partial and ordinary differential equations being a modified version of the Hodgkin-Huxley model and SM fibre mechanics. In numerical experiments; the source of excitation is mechanical stretches of SM at a fixed amplitude and variable frequencies. Results: Low frequency (0.5 < v < 2 Hz) stimuli cause the propagation of spikes in the neuronal chain and, finally, the generation of active forces by SM. However, induced contractions are not sufficient to initiate travelling wave dynamics in the control system. At frequencies, 2 < v < 4 Hz, multiple low amplitude and short-lasting contractions are observed in SM after the termination of stretching. For frequencies (0.5 < v < 4 Hz), primary and sensory neurones demonstrate strong connectivity and coherent electrical activity. Significant qualitative and quantitative changes in dynamics of myoelectical patterns with a transition to a self-organised mode are recorded with the high degree of stretches at v = 4.5 Hz. Increased rates of deformation lead to the production of high amplitude signals at the mechanoreceptors with subsequent self-sustained excitation within the neuronal chain. Remarkably, the connection between neurones weakens resulting in incoherent firing. Further increase in a frequency of stimulation (v > 4.5 Hz) has a detrimental effect on the system. The mechanical and control systems become disconnected and exhibit uncoordinated electromechanical activity. Conclusion: To our knowledge, the existence of periodic activity in a multicellular, functionally heterogeneous biological system with mechano-electrical dynamics, such as the FU, has been demonstrated for the first time. These findings support the notion of possible peristalsis in the gut even in the absence of intrinsic sources - pacemaker cells. Results could be implicated in the pathogenesis of intestinal dysrythmia, a medical condition associated with motor dysfunction.

Keywords: complex dynamics, functional unit, the gut, dysrythmia

Procedia PDF Downloads 202
1129 Detection of Polymorphism of Growth Hormone Gene in Holstein Cattle

Authors: Emine Şahin, Murat Soner Balcıoğlu

Abstract:

The aim of this study was to determine the growth hormone (bGH) gene polymorphism in the Holstein cattle growing around Antalya in Turkey. In order to determine the bGH-AluI polymorphism, polymerase chain reaction - restriction fragment length polymorphism (PCR-RFLP) method was performed. A 891 bp fragment of bGH was amplified and two types of alleles C and D for bGH were observed. In this study, the frequencies of C and D alleles were 0.8438 and 0.1562, respectively. The genotype frequencies for CC, CD and DD were 0.787, 0.191 and 0.022, respectively. According to the results of the chi-square test, a significant deviation from the Hardy-Weinberg equilibrium was not determined for the bGH locus in the population.

Keywords: Growth Hormone Gene, Holstein , Polymorphism, RFLP

Procedia PDF Downloads 367
1128 The Need for Automation in the Domestic Food Processing Sector and its Impact

Authors: Shantam Gupta

Abstract:

The objective of this study is to address the critical need for automation in the domestic food processing sector and study its impact. Food is the one of the most basic physiological needs essential for the survival of a living being. Some of them have the capacity to prepare their own food (like most plants) and henceforth are designated as primary food producers; those who depend on these primary food producers for food form the primary consumers’ class (herbivores). Some of the organisms relying on the primary food are the secondary food consumers (carnivores). There is a third class of consumers called tertiary food consumers/apex food consumers that feed on both the primary and secondary food consumers. Humans form an essential part of the apex predators and are generally at the top of the food chain. But still further disintegration of the food habits of the modern human i.e. Homo sapiens, reveals that humans depend on other individuals for preparing their own food. The old notion of eating raw/brute food is long gone and food processing has become very trenchant in lives of modern human. This has led to an increase in dependence on other individuals for ‘processing’ the food before it can be actually consumed by the modern human. This has led to a further shift of humans in the classification of food chain of consumers. The effects of the shifts shall be systematically investigated in this paper. The processing of food has a direct impact on the economy of the individual (consumer). Also most individuals depend on other processing individuals for the preparation of food. This dependency leads to establishment of a vital link of dependency in the food web which when altered can adversely affect the food web and can have dire consequences on the health of the individual. This study investigates the challenges arising out due to this dependency and the impact of food processing on the economy of the individual. A comparison of Industrial food processing and processing at domestic platforms (households and restaurants) has been made to provide an idea about the present scenario of automation in the food processing sector. A lot of time and energy is also consumed while processing food at home for consumption. The high frequency of consumption of meals (greater than 2 times a day) makes it even more laborious. Through the medium of this study a pressing need for development of an automatic cooking machine is proposed with a mission to reduce the inter-dependency & human effort of individuals required for the preparation of food (by automation of the food preparation process) and make them more self-reliant The impact of development of this product has also further been profoundly discussed. Assumption used: The individuals those who process food also consume the food that they produce. (They are also termed as ‘independent’ or ‘self-reliant’ modern human beings.)

Keywords: automation, food processing, impact on economy, processing individual

Procedia PDF Downloads 466
1127 Quantitative Polymerase Chain Reaction Analysis of Phytoplankton Composition and Abundance to Assess Eutrophication: A Multi-Year Study in Twelve Large Rivers across the United States

Authors: Chiqian Zhang, Kyle D. McIntosh, Nathan Sienkiewicz, Ian Struewing, Erin A. Stelzer, Jennifer L. Graham, Jingrang Lu

Abstract:

Phytoplankton plays an essential role in freshwater aquatic ecosystems and is the primary group synthesizing organic carbon and providing food sources or energy to ecosystems. Therefore, the identification and quantification of phytoplankton are important for estimating and assessing ecosystem productivity (carbon fixation), water quality, and eutrophication. Microscopy is the current gold standard for identifying and quantifying phytoplankton composition and abundance. However, microscopic analysis of phytoplankton is time-consuming, has a low sample throughput, and requires deep knowledge and rich experience in microbial morphology to implement. To improve this situation, quantitative polymerase chain reaction (qPCR) was considered for phytoplankton identification and quantification. Using qPCR to assess phytoplankton composition and abundance, however, has not been comprehensively evaluated. This study focused on: 1) conducting a comprehensive performance comparison of qPCR and microscopy techniques in identifying and quantifying phytoplankton and 2) examining the use of qPCR as a tool for assessing eutrophication. Twelve large rivers located throughout the United States were evaluated using data collected from 2017 to 2019 to understand the relation between qPCR-based phytoplankton abundance and eutrophication. This study revealed that temporal variation of phytoplankton abundance in the twelve rivers was limited within years (from late spring to late fall) and among different years (2017, 2018, and 2019). Midcontinent rivers had moderately greater phytoplankton abundance than eastern and western rivers, presumably because midcontinent rivers were more eutrophic. The study also showed that qPCR- and microscope-determined phytoplankton abundance had a significant positive linear correlation (adjusted R² 0.772, p-value < 0.001). In addition, phytoplankton abundance assessed via qPCR showed promise as an indicator of the eutrophication status of those rivers, with oligotrophic rivers having low phytoplankton abundance and eutrophic rivers having (relatively) high phytoplankton abundance. This study demonstrated that qPCR could serve as an alternative tool to traditional microscopy for phytoplankton quantification and eutrophication assessment in freshwater rivers.

Keywords: phytoplankton, eutrophication, river, qPCR, microscopy, spatiotemporal variation

Procedia PDF Downloads 97
1126 A Machine Learning Approach for Classification of Directional Valve Leakage in the Hydraulic Final Test

Authors: Christian Neunzig, Simon Fahle, Jürgen Schulz, Matthias Möller, Bernd Kuhlenkötter

Abstract:

Due to increasing cost pressure in global markets, artificial intelligence is becoming a technology that is decisive for competition. Predictive quality enables machinery and plant manufacturers to ensure product quality by using data-driven forecasts via machine learning models as a decision-making basis for test results. The use of cross-process Bosch production data along the value chain of hydraulic valves is a promising approach to classifying the quality characteristics of workpieces.

Keywords: predictive quality, hydraulics, machine learning, classification, supervised learning

Procedia PDF Downloads 226
1125 Modelling Urban Rigidity and Elasticity Growth Boundaries: A Spatial Constraints-Suitability Based Perspective

Authors: Pengcheng Xiang Jr., Xueqing Sun, Dong Ngoduy

Abstract:

In the context of rapid urbanization, urban sprawl has brought about extensive negative impacts on ecosystems and the environment, resulting in a gradual shift from "incremental growth" to ‘stock growth’ in cities. A detailed urban growth boundary is a prerequisite for urban renewal and management. This study takes Shenyang City, China, as the study area and evaluates the spatial distribution of urban spatial suitability in the study area from the perspective of spatial constraints-suitability using multi-source data and simulates the future rigid and elastic growth boundaries of the city in the study area using the CA-Markov model. The results show that (1) the suitable construction area and moderate construction area in the study area account for 8.76% and 19.01% of the total area, respectively, and the suitable construction area and moderate construction area show a trend of distribution from the urban centre to the periphery, mainly in Shenhe District, the southern part of Heping District, the western part of Dongling District, and the central part of Dadong District; (2) the area of expansion of construction land in the study area in the period of 2023-2030 is 153274.6977hm2, accounting for 44.39% of the total area of the study area; (3) the rigid boundary of the study area occupies an area of 153274.6977 hm2, accounting for 44.39% of the total area of the study area, and the elastic boundary of the study area contains an area of 75362.61 hm2, accounting for 21.69% of the total area of the study area. The study constructed a method for urban growth boundary delineation, which helps to apply remote sensing to guide future urban spatial growth management and urban renewal.

Keywords: urban growth boundary, spatial constraints, spatial suitability, urban sprawl

Procedia PDF Downloads 28
1124 Incorporation of Noncanonical Amino Acids into Hard-to-Express Antibody Fragments: Expression and Characterization

Authors: Hana Hanaee-Ahvaz, Monika Cserjan-Puschmann, Christopher Tauer, Gerald Striedner

Abstract:

Incorporation of noncanonical amino acids (ncAA) into proteins has become an interesting topic as proteins featured with ncAAs offer a wide range of different applications. Nowadays, technologies and systems exist that allow for the site-specific introduction of ncAAs in vivo, but the efficient production of proteins modified this way is still a big challenge. This is especially true for 'hard-to-express' proteins where low yields are encountered even with the native sequence. In this study, site-specific incorporation of azido-ethoxy-carbonyl-Lysin (azk) into an anti-tumor-necrosis-factor-α-Fab (FTN2) was investigated. According to well-established parameters, possible site positions for ncAA incorporation were determined, and corresponding FTN2 genes were constructed. Each of the modified FTN2 variants has one amber codon for azk incorporated either in its heavy or light chain. The expression level for all variants produced was determined by ELISA, and all azk variants could be produced with a satisfactory yield in the range of 50-70% of the original FTN2 variant. In terms of expression yield, neither the azk incorporation position nor the subunit modified (heavy or light chain) had a significant effect. We confirmed correct protein processing and azk incorporation by mass spectrometry analysis, and antigen-antibody interaction was determined by surface plasmon resonance analysis. The next step is to characterize the effect of azk incorporation on protein stability and aggregation tendency via differential scanning calorimetry and light scattering, respectively. In summary, the incorporation of ncAA into our Fab candidate FTN2 worked better than expected. The quantities produced allowed a detailed characterization of the variants in terms of their properties, and we can now turn our attention to potential applications. By using click chemistry, we can equip the Fabs with additional functionalities and make them suitable for a wide range of applications. We will now use this option in a first approach and develop an assay that will allow us to follow the degradation of the recombinant target protein in vivo. Special focus will be laid on the proteolytic activity in the periplasm and how it is influenced by cultivation/induction conditions.

Keywords: degradation, FTN2, hard-to-express protein, non-canonical amino acids

Procedia PDF Downloads 229
1123 A Decision-Support Tool for Humanitarian Distribution Planners in the Face of Congestion at Security Checkpoints: A Real-World Case Study

Authors: Mohanad Rezeq, Tarik Aouam, Frederik Gailly

Abstract:

In times of armed conflicts, various security checkpoints are placed by authorities to control the flow of merchandise into and within areas of conflict. The flow of humanitarian trucks that is added to the regular flow of commercial trucks, together with the complex security procedures, creates congestion and long waiting times at the security checkpoints. This causes distribution costs to increase and shortages of relief aid to the affected people to occur. Our research proposes a decision-support tool to assist planners and policymakers in building efficient plans for the distribution of relief aid, taking into account congestion at security checkpoints. The proposed tool is built around a multi-item humanitarian distribution planning model based on multi-phase design science methodology that has as its objective to minimize distribution and back ordering costs subject to capacity constraints that reflect congestion effects using nonlinear clearing functions. Using the 2014 Gaza War as a case study, we illustrate the application of the proposed tool, model the underlying relief-aid humanitarian supply chain, estimate clearing functions at different security checkpoints, and conduct computational experiments. The decision support tool generated a shipment plan that was compared to two benchmarks in terms of total distribution cost, average lead time and work in progress (WIP) at security checkpoints, and average inventory and backorders at distribution centers. The first benchmark is the shipment plan generated by the fixed capacity model, and the second is the actual shipment plan implemented by the planners during the armed conflict. According to our findings, modeling and optimizing supply chain flows reduce total distribution costs, average truck wait times at security checkpoints, and average backorders when compared to the executed plan and the fixed-capacity model. Finally, scenario analysis concludes that increasing capacity at security checkpoints can lower total operations costs by reducing the average lead time.

Keywords: humanitarian distribution planning, relief-aid distribution, congestion, clearing functions

Procedia PDF Downloads 81
1122 A Two-Step, Temperature-Staged, Direct Coal Liquefaction Process

Authors: Reyna Singh, David Lokhat, Milan Carsky

Abstract:

The world crude oil demand is projected to rise to 108.5 million bbl/d by the year 2035. With reserves estimated at 869 billion tonnes worldwide, coal is an abundant resource. This work was aimed at producing a high value hydrocarbon liquid product from the Direct Coal Liquefaction (DCL) process at, comparatively, mild operating conditions. Via hydrogenation, the temperature-staged approach was investigated. In a two reactor lab-scale pilot plant facility, the objectives included maximising thermal dissolution of the coal in the presence of a hydrogen donor solvent in the first stage, subsequently promoting hydrogen saturation and hydrodesulphurization (HDS) performance in the second. The feed slurry consisted of high grade, pulverized bituminous coal on a moisture-free basis with a size fraction of < 100μm; and Tetralin mixed in 2:1 and 3:1 solvent/coal ratios. Magnetite (Fe3O4) at 0.25wt% of the dry coal feed was added for the catalysed runs. For both stages, hydrogen gas was used to maintain a system pressure of 100barg. In the first stage, temperatures of 250℃ and 300℃, reaction times of 30 and 60 minutes were investigated in an agitated batch reactor. The first stage liquid product was pumped into the second stage vertical reactor, which was designed to counter-currently contact the hydrogen rich gas stream and incoming liquid flow in the fixed catalyst bed. Two commercial hydrotreating catalysts; Cobalt-Molybdenum (CoMo) and Nickel-Molybdenum (NiMo); were compared in terms of their conversion, selectivity and HDS performance at temperatures 50℃ higher than the respective first stage tests. The catalysts were activated at 300°C with a hydrogen flowrate of approximately 10 ml/min prior to the testing. A gas-liquid separator at the outlet of the reactor ensured that the gas was exhausted to the online VARIOplus gas analyser. The liquid was collected and sampled for analysis using Gas Chromatography-Mass Spectrometry (GC-MS). Internal standard quantification methods for the sulphur content, the BTX (benzene, toluene, and xylene) and alkene quality; alkanes and polycyclic aromatic hydrocarbon (PAH) compounds in the liquid products were guided by ASTM standards of practice for hydrocarbon analysis. In the first stage, using a 2:1 solvent/coal ratio, an increased coal to liquid conversion was favoured by a lower operating temperature of 250℃, 60 minutes and a system catalysed by magnetite. Tetralin functioned effectively as the hydrogen donor solvent. A 3:1 ratio favoured increased concentrations of the long chain alkanes undecane and dodecane, unsaturated alkenes octene and nonene and PAH compounds such as indene. The second stage product distribution showed an increase in the BTX quality of the liquid product, branched chain alkanes and a reduction in the sulphur concentration. As an HDS performer and selectivity to the production of long and branched chain alkanes, NiMo performed better than CoMo. CoMo is selective to a higher concentration of cyclohexane. For 16 days on stream each, NiMo had a higher activity than CoMo. The potential to cover the demand for low–sulphur, crude diesel and solvents from the production of high value hydrocarbon liquid in the said process, is thus demonstrated.

Keywords: catalyst, coal, liquefaction, temperature-staged

Procedia PDF Downloads 644
1121 Evaluation of Main Factors Affecting the Choice of a Freight Forwarder: A Sri Lankan Exporter’s Perspective

Authors: Ishani Maheshika

Abstract:

The intermediary role performed by freight forwarders in exportation has become significant in fulfilling businesses’ supply chain needs in this dynamic world. Since the success of exporter’s business is at present, highly reliant on supply chain optimization, cost efficiency, profitability, consistent service and responsiveness, the decision of selecting the most beneficial freight forwarder has become crucial for exporters. Although there are similar foreign researches, prior researches covering Sri Lankan setting are not in existence. Moreover, results vary with time, nature of industry and business environment factors. Therefore, a study from the perspective of Sri Lankan exporters was identified as a requisite to be researched. In order to identify and prioritize key factors which have affected the exporter’s decision in selecting freight forwarders in Sri Lankan context, Sri Lankan export industry was stratified into 22 sectors based on commodity using stratified sampling technique. One exporter from each sector was then selected using judgmental sampling to have a sample of 22. Factors which were identified through a pilot survey, was organized under 6 main criteria. A questionnaire was basically developed as pairwise comparisons using 9-point semantic differential scale and comparisons were done within main criteria and subcriteria. After a pre-testing, interviews and e-mail questionnaire survey were conducted. Data were analyzed using Analytic Hierarchy Process to determine priority vectors of criteria. Customer service was found to be the most important main criterion for Sri Lankan exporters. It was followed by reliability and operational efficiency respectively. The criterion of the least importance is company background and reputation. Whereas small sized exporters pay more attention to rate, reliability is the major concern among medium and large scale exporters. Irrespective of seniority of the exporter, reliability is given the prominence. Responsiveness is the most important sub criterion among Sri Lankan exporters. Consistency of judgments with respect to main criteria was verified through consistency ratio, which was less than 10%. Being more competitive, freight forwarders should come up with customized marketing strategies based on each target group’s requirements and expectations in offering services to retain existing exporters and attract new exporters.

Keywords: analytic hierarchy process, freight forwarders, main criteria, Sri Lankan exporters, subcriteria

Procedia PDF Downloads 403
1120 A Review on Artificial Neural Networks in Image Processing

Authors: B. Afsharipoor, E. Nazemi

Abstract:

Artificial neural networks (ANNs) are powerful tool for prediction which can be trained based on a set of examples and thus, it would be useful for nonlinear image processing. The present paper reviews several paper regarding applications of ANN in image processing to shed the light on advantage and disadvantage of ANNs in this field. Different steps in the image processing chain including pre-processing, enhancement, segmentation, object recognition, image understanding and optimization by using ANN are summarized. Furthermore, results on using multi artificial neural networks are presented.

Keywords: neural networks, image processing, segmentation, object recognition, image understanding, optimization, MANN

Procedia PDF Downloads 402
1119 Exploring Managerial Approaches towards Green Manufacturing: A Thematic Analysis

Authors: Hakimeh Masoudigavgani

Abstract:

Since manufacturing firms deplete non-renewable resources and pollute air, soil, and water in greatly unsustainable manner, industrial activities or production of products are considered to be a key contributor to adverse environmental impacts. Hence, management strategies and approaches that involve an effective supply chain decision process in a manufacturing sector could be extremely significant to the application of environmental initiatives. Green manufacturing (GM) is one of these strategies which minimises negative effects on the environment through reducing greenhouse gas emissions, waste, and the consumption of energy and natural resources. This paper aims to explore what greening methods and mechanisms could be applied in the manufacturing supply chain and what are the outcomes of adopting these methods in terms of abating environmental burdens? The study is an interpretive research with an exploratory approach, using thematic analysis by coding text, breaking down and grouping the content of collected literature into various themes and categories. It is found that green supply chain could be attained through execution of some pre-production strategies including green building, eco-design, and green procurement as well as a number of in-production and post-production strategies involving green manufacturing and green logistics. To achieve an effective GM, the pre-production strategies are suggested to be employed. This paper defines GM as (1) the analysis of the ecological impacts generated by practices, products, production processes, and operational functions, and (2) the implementation of greening methods to reduce damaging influences of them on the natural environment. Analysis means assessing, monitoring, and auditing of practices in order to measure and pinpoint their harmful impacts. Moreover, greening methods involved within GM (arranged in order from the least to the most level of environmental compliance and techniques) consist of: •product stewardship (e.g. less use of toxic, non-renewable, and hazardous materials in the manufacture of the product; and stewardship of the environmental problems with regard to the product in all production, use, and end-of-life stages); •process stewardship (e.g. controlling carbon emission, energy and resources usage, transportation method, and disposal; reengineering polluting processes; recycling waste materials generated in production); •lean and clean production practices (e.g. elimination of waste, materials replacement, materials reduction, resource-efficient consumption, energy-efficient usage, emission reduction, managerial assessment, waste re-use); •use of eco-industrial parks (e.g. a shared warehouse, shared logistics management system, energy co-generation plant, effluent treatment). However, the focus of this paper is only on methods related to the in-production phase and needs further research on both pre-production and post-production environmental innovations. The outlined methods in this investigation may possibly be taken into account by policy/decision makers. Additionally, the proposed future research direction and identified gaps can be filled by scholars and researchers. The paper compares and contrasts a variety of viewpoints and enhances the body of knowledge by building a definition for GM through synthesising literature and categorising the strategic concept of greening methods, drivers, barriers, and successful implementing tactics.

Keywords: green manufacturing (GM), product stewardship, process stewardship, clean production, eco-industrial parks (EIPs)

Procedia PDF Downloads 579
1118 Theoretical Modelling of Molecular Mechanisms in Stimuli-Responsive Polymers

Authors: Catherine Vasnetsov, Victor Vasnetsov

Abstract:

Context: Thermo-responsive polymers are materials that undergo significant changes in their physical properties in response to temperature changes. These polymers have gained significant attention in research due to their potential applications in various industries and medicine. However, the molecular mechanisms underlying their behavior are not well understood, particularly in relation to cosolvency, which is crucial for practical applications. Research Aim: This study aimed to theoretically investigate the phenomenon of cosolvency in long-chain polymers using the Flory-Huggins statistical-mechanical framework. The main objective was to understand the interactions between the polymer, solvent, and cosolvent under different conditions. Methodology: The research employed a combination of Monte Carlo computer simulations and advanced machine-learning methods. The Flory-Huggins mean field theory was used as the basis for the simulations. Spinodal graphs and ternary plots were utilized to develop an initial computer model for predicting polymer behavior. Molecular dynamic simulations were conducted to mimic real-life polymer systems. Machine learning techniques were incorporated to enhance the accuracy and reliability of the simulations. Findings: The simulations revealed that the addition of very low or very high volumes of cosolvent molecules resulted in smaller radii of gyration for the polymer, indicating poor miscibility. However, intermediate volume fractions of cosolvent led to higher radii of gyration, suggesting improved miscibility. These findings provide a possible microscopic explanation for the cosolvency phenomenon in polymer systems. Theoretical Importance: This research contributes to a better understanding of the behavior of thermo-responsive polymers and the role of cosolvency. The findings provide insights into the molecular mechanisms underlying cosolvency and offer specific predictions for future experimental investigations. The study also presents a more rigorous analysis of the Flory-Huggins free energy theory in the context of polymer systems. Data Collection and Analysis Procedures: The data for this study was collected through Monte Carlo computer simulations and molecular dynamic simulations. The interactions between the polymer, solvent, and cosolvent were analyzed using the Flory-Huggins mean field theory. Machine learning techniques were employed to enhance the accuracy of the simulations. The collected data was then analyzed to determine the impact of cosolvent volume fractions on the radii of gyration of the polymer. Question Addressed: The research addressed the question of how cosolvency affects the behavior of long-chain polymers. Specifically, the study aimed to investigate the interactions between the polymer, solvent, and cosolvent under different volume fractions and understand the resulting changes in the radii of gyration. Conclusion: In conclusion, this study utilized theoretical modeling and computer simulations to investigate the phenomenon of cosolvency in long-chain polymers. The findings suggest that moderate cosolvent volume fractions can lead to improved miscibility, as indicated by higher radii of gyration. These insights contribute to a better understanding of the molecular mechanisms underlying cosolvency in polymer systems and provide predictions for future experimental studies. The research also enhances the theoretical analysis of the Flory-Huggins free energy theory.

Keywords: molecular modelling, flory-huggins, cosolvency, stimuli-responsive polymers

Procedia PDF Downloads 68
1117 Solving Extended Linear Complementarity Problems (XLCP) - Wood and Environment

Authors: Liberto Pombal, Christian Dieter Jaekel

Abstract:

The objective of this work is to establish theoretical and numerical conditions for Solving Extended Linear Complementarity Problems (XLCP), with emphasis on the Horizontal Linear Complementarity Problem (HLCP). Two new strategies for solving complementarity problems are presented, using differentiable and penalized functions, which resulted in a natural formalization for the Linear Horizontal case. The computational results of all suggested strategies are also discussed in depth in this paper. The implication in practice allows solving and optimizing, in an innovative way, the (forestry) problems of the value chain of the industrial wood sector in Angola.

Keywords: complementarity, box constrained, optimality conditions, wood and environment

Procedia PDF Downloads 52
1116 Deep Reinforcement Learning-Based Computation Offloading for 5G Vehicle-Aware Multi-Access Edge Computing Network

Authors: Ziying Wu, Danfeng Yan

Abstract:

Multi-Access Edge Computing (MEC) is one of the key technologies of the future 5G network. By deploying edge computing centers at the edge of wireless access network, the computation tasks can be offloaded to edge servers rather than the remote cloud server to meet the requirements of 5G low-latency and high-reliability application scenarios. Meanwhile, with the development of IOV (Internet of Vehicles) technology, various delay-sensitive and compute-intensive in-vehicle applications continue to appear. Compared with traditional internet business, these computation tasks have higher processing priority and lower delay requirements. In this paper, we design a 5G-based Vehicle-Aware Multi-Access Edge Computing Network (VAMECN) and propose a joint optimization problem of minimizing total system cost. In view of the problem, a deep reinforcement learning-based joint computation offloading and task migration optimization (JCOTM) algorithm is proposed, considering the influences of multiple factors such as concurrent multiple computation tasks, system computing resources distribution, and network communication bandwidth. And, the mixed integer nonlinear programming problem is described as a Markov Decision Process. Experiments show that our proposed algorithm can effectively reduce task processing delay and equipment energy consumption, optimize computing offloading and resource allocation schemes, and improve system resource utilization, compared with other computing offloading policies.

Keywords: multi-access edge computing, computation offloading, 5th generation, vehicle-aware, deep reinforcement learning, deep q-network

Procedia PDF Downloads 114
1115 Multiscale Modelling of Citrus Black Spot Transmission Dynamics along the Pre-Harvest Supply Chain

Authors: Muleya Nqobile, Winston Garira

Abstract:

We presented a compartmental deterministic multi-scale model which encompass internal plant defensive mechanism and pathogen interaction, then we consider nesting the model into the epidemiological model. The objective was to improve our understanding of the transmission dynamics of within host and between host of Guignardia citricapa Kiely. The inflow of infected class was scaled down to individual level while the outflow was scaled up to average population level. Conceptual model and mathematical model were constructed to display a theoretical framework which can be used for predicting or identify disease pattern.

Keywords: epidemiological model, mathematical modelling, multi-scale modelling, immunological model

Procedia PDF Downloads 455
1114 Fast Bayesian Inference of Multivariate Block-Nearest Neighbor Gaussian Process (NNGP) Models for Large Data

Authors: Carlos Gonzales, Zaida Quiroz, Marcos Prates

Abstract:

Several spatial variables collected at the same location that share a common spatial distribution can be modeled simultaneously through a multivariate geostatistical model that takes into account the correlation between these variables and the spatial autocorrelation. The main goal of this model is to perform spatial prediction of these variables in the region of study. Here we focus on a geostatistical multivariate formulation that relies on sharing common spatial random effect terms. In particular, the first response variable can be modeled by a mean that incorporates a shared random spatial effect, while the other response variables depend on this shared spatial term, in addition to specific random spatial effects. Each spatial random effect is defined through a Gaussian process with a valid covariance function, but in order to improve the computational efficiency when the data are large, each Gaussian process is approximated to a Gaussian random Markov field (GRMF), specifically to the block nearest neighbor Gaussian process (Block-NNGP). This approach involves dividing the spatial domain into several dependent blocks under certain constraints, where the cross blocks allow capturing the spatial dependence on a large scale, while each individual block captures the spatial dependence on a smaller scale. The multivariate geostatistical model belongs to the class of Latent Gaussian Models; thus, to achieve fast Bayesian inference, it is used the integrated nested Laplace approximation (INLA) method. The good performance of the proposed model is shown through simulations and applications for massive data.

Keywords: Block-NNGP, geostatistics, gaussian process, GRMF, INLA, multivariate models.

Procedia PDF Downloads 93
1113 High Purity Lignin for Asphalt Applications: Using the Dawn Technology™ Wood Fractionation Process

Authors: Ed de Jong

Abstract:

Avantium is a leading technology development company and a frontrunner in renewable chemistry. Avantium develops disruptive technologies that enable the production of sustainable high value products from renewable materials and actively seek out collaborations and partnerships with like-minded companies and academic institutions globally, to speed up introductions of chemical innovations in the marketplace. In addition, Avantium helps companies to accelerate their catalysis R&D to improve efficiencies and deliver increased sustainability, growth, and profits, by providing proprietary systems and services to this regard. Many chemical building blocks and materials can be produced from biomass, nowadays mainly from 1st generation based carbohydrates, but potential for competition with the human food chain leads brand-owners to look for strategies to transition from 1st to 2nd generation feedstock. The use of non-edible lignocellulosic feedstock is an equally attractive source to produce chemical intermediates and an important part of the solution addressing these global issues (Paris targets). Avantium’s Dawn Technology™ separates the glucose, mixed sugars, and lignin available in non-food agricultural and forestry residues such as wood chips, wheat straw, bagasse, empty fruit bunches or corn stover. The resulting very pure lignin is dense in energy and can be used for energy generation. However, such a material might preferably be deployed in higher added value applications. Bitumen, which is fossil based, are mostly used for paving applications. Traditional hot mix asphalt emits large quantities of the GHG’s CO₂, CH₄, and N₂O, which is unfavorable for obvious environmental reasons. Another challenge for the bitumen industry is that the petrochemical industry is becoming more and more efficient in breaking down higher chain hydrocarbons to lower chain hydrocarbons with higher added value than bitumen. This has a negative effect on the availability of bitumen. The asphalt market, as well as governments, are looking for alternatives with higher sustainability in terms of GHG emission. The usage of alternative sustainable binders, which can (partly) replace the bitumen, contributes to reduce GHG emissions and at the same time broadens the availability of binders. As lignin is a major component (around 25-30%) of lignocellulosic material, which includes terrestrial plants (e.g., trees, bushes, and grass) and agricultural residues (e.g., empty fruit bunches, corn stover, sugarcane bagasse, straw, etc.), it is globally highly available. The chemical structure shows resemblance with the structure of bitumen and could, therefore, be used as an alternative for bitumen in applications like roofing or asphalt. Applications such as the use of lignin in asphalt need both fundamental research as well as practical proof under relevant use conditions. From a fundamental point of view, rheological aspects, as well as mixing, are key criteria. From a practical point of view, behavior in real road conditions is key (how easy can the asphalt be prepared, how easy can it be applied on the road, what is the durability, etc.). The paper will discuss the fundamentals of the use of lignin as bitumen replacement as well as the status of the different demonstration projects in Europe using lignin as a partial bitumen replacement in asphalts and will especially present the results of using Dawn Technology™ lignin as partial replacement of bitumen.

Keywords: biorefinery, wood fractionation, lignin, asphalt, bitumen, sustainability

Procedia PDF Downloads 152
1112 An Optimization Modelling to Evaluate Flights Scheduling at Tourist Airports

Authors: Dimitrios J. Dimitriou

Abstract:

Airport’s serving a tourist destination are an essential counterpart of the tourist demand supply chain, and their productivity is related to the region’s attractiveness and is enhanced by the air transport business. In this paper, the evaluation framework of the scheduled flights between two tourist airports is taken into consideration. By adopting a systemic approach, the arrivals from an airport that its connectivity heavily depended on the departures of another major airport are reviewed. The methodology framework, based on inventory control theory and the numerical example, promotes the use of the modelling formulation. The results would be essential for comparison and exercising to other similar cases.

Keywords: airport connectivity, inventory control, optimization, optimum allocation

Procedia PDF Downloads 331
1111 Genotyping of Rotaviruses in Pediatric Patients with Gastroenteritis by Using Real-Time Reverse Transcription Polymerase Chain Reaction

Authors: Recep Kesli, Cengiz Demir, Riza Durmaz, Zekiye Bakkaloglu, Aysegul Bukulmez

Abstract:

Objective: Acute diarrhea disease in children is a major cause of morbidity worldwide and is a leading cause of mortality, and it is the most common agent responsible for acute gastroenteritis in developing countries. With hospitalized children suffering from acute enteric disease up to 50% of the analyzed specimen were positive for rotavirus. Further molecular surveillance could provide a sound basis for improving the response to epidemic gastroenteritis and could provide data needed for the introduction of vaccination programmes in the country. The aim of this study was to investigate the prevalence of viral etiology of the gastroenteritis in children aged 0-6 years with acute gastroenteritis and to determine predominant genotypes of rotaviruses in the province of Afyonkarahisar, Turkey. Methods: An epidemiological study on rotavirus was carried out during 2016. Fecal samples obtained from the 144 rotavirus positive children with 0-6 years of ages and applied to the Pediatric Diseases Outpatient of ANS Research and Practice Hospital, Afyon Kocatepe University with the complaint of diarrhea. Bacterial agents causing gastroenteritis were excluded by using bacteriological culture methods and finally, no growth observed. Rotavirus antigen was examined by both the immunochromatographic (One Step Rotavirus and Adenovirus Combo Test, China) and ELISA (Premier Rotaclone, USA) methods in stool samples. Rotavirus RNA was detected by using one step real-time reverse transcription-polymerase chain reaction (RT-PCR). G and P genotypes were determined using RT-PCR with consensus primers of VP7 and VP4 genes, followed by semi nested type-specific multiplex PCR. Results: Of the total 144 rotavirus antigen-positive samples with RT-PCR, 4 (2,8%) were rejected, 95 (66%) were examined, and 45 (31,2%) have not been examined for PCR yet. Ninety-one (95,8%) of the 95 examined samples were found to be rotavirus positive with RT-PCR. Rotavirus subgenotyping distributions in G, P and G/P genotype groups were determined as; G1:45%, G2:27%, G3:13%, G9:13%, G4:1% and G12:1% for G genotype, and P[4]:33%, P[8]:66%, P[10]:1% for P genotype, and G1P[8]:%37, G2P[4]:%21, G3P[8]:%10, G4P[8]:%1, G9P[8]:%8, G2P[8]:%3 for G/P genotype . Not common genotype combination were %20 in G/P genotype. Conclusions: This study subscribes to the global agreement of the molecular epidemiology of rotavirus which will be useful in guiding the alternative and application of rotavirus vaccines or effective control and interception. Determining the diversity and rates of rotavirus genotypes will definitely provide guidelines for developing the most suitable vaccine.

Keywords: gastroenteritis, genotyping, rotavirus, RT-PCR

Procedia PDF Downloads 234
1110 Predicting Aggregation Propensity from Low-Temperature Conformational Fluctuations

Authors: Hamza Javar Magnier, Robin Curtis

Abstract:

There have been rapid advances in the upstream processing of protein therapeutics, which has shifted the bottleneck to downstream purification and formulation. Finding liquid formulations with shelf lives of up to two years is increasingly difficult for some of the newer therapeutics, which have been engineered for activity, but their formulations are often viscous, can phase separate, and have a high propensity for irreversible aggregation1. We explore means to develop improved predictive ability from a better understanding of how protein-protein interactions on formulation conditions (pH, ionic strength, buffer type, presence of excipients) and how these impact upon the initial steps in protein self-association and aggregation. In this work, we study the initial steps in the aggregation pathways using a minimal protein model based on square-well potentials and discontinuous molecular dynamics. The effect of model parameters, including range of interaction, stiffness, chain length, and chain sequence, implies that protein models fold according to various pathways. By reducing the range of interactions, the folding- and collapse- transition come together, and follow a single-step folding pathway from the denatured to the native state2. After parameterizing the model interaction-parameters, we developed an understanding of low-temperature conformational properties and fluctuations, and the correlation to the folding transition of proteins in isolation. The model fluctuations increase with temperature. We observe a low-temperature point, below which large fluctuations are frozen out. This implies that fluctuations at low-temperature can be correlated to the folding transition at the melting temperature. Because proteins “breath” at low temperatures, defining a native-state as a single structure with conserved contacts and a fixed three-dimensional structure is misleading. Rather, we introduce a new definition of a native-state ensemble based on our understanding of the core conservation, which takes into account the native fluctuations at low temperatures. This approach permits the study of a large range of length and time scales needed to link the molecular interactions to the macroscopically observed behaviour. In addition, these models studied are parameterized by fitting to experimentally observed protein-protein interactions characterized in terms of osmotic second virial coefficients.

Keywords: protein folding, native-ensemble, conformational fluctuation, aggregation

Procedia PDF Downloads 358
1109 Efficient Design of Distribution Logistics by Using a Model-Based Decision Support System

Authors: J. Becker, R. Arnold

Abstract:

The design of distribution logistics has a decisive impact on a company's logistics costs and performance. Hence, such solutions make an essential contribution to corporate success. This article describes a decision support system for analyzing the potential of distribution logistics in terms of logistics costs and performance. In contrast to previous procedures of business process re-engineering (BPR), this method maps distribution logistics holistically under variable distribution structures. Combined with qualitative measures the decision support system will contribute to a more efficient design of distribution logistics.

Keywords: decision support system, distribution logistics, potential analyses, supply chain management

Procedia PDF Downloads 403
1108 Radionuclide Determination Study for Some Fish Species in Kuwait

Authors: Ahmad Almutairi

Abstract:

Kuwait lies to the northwest of the Arabian Gulf. The levels of radionuclides are unknown in this area. Radionuclide like ²¹⁰Po, ²²⁶Ra, and ⁹⁰Sr accumulated in certain body tissues and bones, relate primarily to dietary uptake and inhalation. A large fraction of radiation exposure experienced by individuals comes from food chain transfer. In this study, some types of Kuwait fish were studied for radionuclide determination. These fish were taken from the Kuwaiti water territory during May. The study is to determine the radiation exposure for ²¹⁰Po in some fish species in Kuwait the ²¹⁰Po concentration was found to be between 0.089 and 2.544 Bq/kg the highs was in Zubaidy and the lowest was in Hamour.

Keywords: the radionuclide, radiation exposure, fish species, Zubaida, Hamour

Procedia PDF Downloads 197
1107 Public Procurement and Innovation: A Municipal Approach

Authors: M. Moso-Diez, J. L. Moragues-Oregi, K. Simon-Elorz

Abstract:

Innovation procurement is designed to steer the development of solutions towards concrete public sector needs as a driver for innovation from the demand side (in public services as well as in market opportunities for companies), is horizontally emerging as a new policy instrument. In 2014 the new EU public procurement directives 2014/24/EC and 2014/25/EC reinforced the support for Public Procurement for Innovation, dedicating funding instruments that can be used across all areas supported by Horizon 2020, and targeting potential buyers of innovative solutions: groups of public procurers with similar needs. Under this programme, new policy adapters and networks emerge, aiming to embed innovation criteria into new procurement processes. As these initiatives are in process, research related to is scarce. We argue that Innovation Public Procurement can arise as an innovative policy instrument to public procurement in different policy domains, in spite of existing institutional and cultural barriers (legal guarantee versus innovation). The presentation combines insights from public procurement to supply management chain management in a sustainability and innovation policy arena, as a means of providing understanding of: (1) the circumstances that emerge; (2) the relationship between public and private actors; and (3) the emerging capacities in the definition of the agenda. The policy adopters are the contracting authorities that mainly are at municipal level where they interact with the supply management chain, interconnecting sustainability and climate measures with other policy priorities such as innovation and urban planning; and through the Competitive Dialogue procedure. We found that geography and territory affect both the level of municipal budget (due to municipal income per capita) and its institutional competencies (due to demographic reasons). In spite of the relevance of institutional determinants for public procurement, other factors play an important role such as human factors as well as both public policy and private intervention. The experience is a ‘city project’ (Bilbao) in the field of brownfield decontamination. Brownfield sites typically refer to abandoned or underused industrial and commercial properties—such as old process plants, mining sites, and landfills—that are available but contain low levels of environmental contaminants that may complicate reuse or redevelopment of the land. This article concludes that Innovation Public Procurement in sustainability and climate issues should be further developed both as a policy instrument and as a policy research line that could enable further relevant changes in public procurement as well as in climate innovation.

Keywords: innovation, city projects, public policy, public procurement

Procedia PDF Downloads 307
1106 Densities and Viscosities of Binary Mixture Containing Diethylamine and 2-Alkanol

Authors: Elham jassemi Zargani, Mohammad almasi

Abstract:

Densities and viscosities for binary mixtures of diethylamine + 2 Alkanol (2 propanol up to 2 pentanol) were measured over the entire composition range and temperature interval of 293.15 to 323.15 K. Excess molar volumes V_m^E and viscosity deviations Δη were calculated and correlated by the Redlich−Kister type function to derive the coefficients and estimate the standard error. For mixtures of diethylamine with used 2-alkanols, V_m^E and Δη are negative over the entire range of mole fraction. The observed variations of these parameters, with alkanols chain length and temperature, are discussed in terms of the inter-molecular interactions between the unlike molecules of the binary mixtures.

Keywords: densities, viscosities, diethylamine, 2-alkanol, Redlich-Kister

Procedia PDF Downloads 384
1105 Fallacies of Argumentation in Modern American Political Discourse

Authors: Zarine Avetisyan

Abstract:

The process of speech production and transmission naturally implies the occurrence of certain defective assumptions and erroneous formulations which may be both spontaneous, caused by haste, carelessness, etc., or deliberate. Whether deliberate or not, fallacies always act by way of “faux pas”. In the latter case, we deal with fake or deceptive arguments which are the focus of the given paper. The paper departs from the assumption that fallacies are arguments that prove nothing. Additionally and more importantly, political discourse becomes the main domain for scholarly “cultivation” while pinning down fallacies. The fallacy of telling the truth but deliberately omitting important key details in order to falsify the larger picture called “the half truth” captures special attention in the given paper.

Keywords: break in the information chain, fallacy, half truth, political discourse

Procedia PDF Downloads 387
1104 Primal Instinct: Formation of Food Aversion

Authors: Zihuan (Dylan) Wang

Abstract:

This paper analyzes the formation of human food aversion from a biological perspective. It points out that this biased behavior is formed through the accumulation of long-term survival and life experiences. By introducing the "Food Chain Energy Pyramid" model and the analogous deduction of the "Human Food Aversion Pyramid," with energy conversion efficiency as the primary reason, it analyzes the underlying reasons for the formation of food preferences. Food industry professionals can gain inspiration from this article to combine the theory presented with their expertise in order to leverage product quality and promote environmentally conscious practices.

Keywords: food aversion, food preference, energy conversion efficiency, food and culture, nutrition, research and development

Procedia PDF Downloads 54