Search results for: cloud network
2761 Cracks Detection and Measurement Using VLP-16 LiDAR and Intel Depth Camera D435 in Real-Time
Authors: Xinwen Zhu, Xingguang Li, Sun Yi
Abstract:
Crack is one of the most common damages in buildings, bridges, roads and so on, which may pose safety hazards. However, cracks frequently happen in structures of various materials. Traditional methods of manual detection and measurement, which are known as subjective, time-consuming, and labor-intensive, are gradually unable to meet the needs of modern development. In addition, crack detection and measurement need be safe considering space limitations and danger. Intelligent crack detection has become necessary research. In this paper, an efficient method for crack detection and quantification using a 3D sensor, LiDAR, and depth camera is proposed. This method works even in a dark environment, which is usual in real-world applications. The LiDAR rapidly spins to scan the surrounding environment and discover cracks through lasers thousands of times per second, providing a rich, 3D point cloud in real-time. The LiDAR provides quite accurate depth information. The precision of the distance of each point can be determined within around ±3 cm accuracy, and not only it is good for getting a precise distance, but it also allows us to see far of over 100m going with the top range models. But the accuracy is still large for some high precision structures of material. To make the depth of crack is much more accurate, the depth camera is in need. The cracks are scanned by the depth camera at the same time. Finally, all data from LiDAR and Depth cameras are analyzed, and the size of the cracks can be quantified successfully. The comparison shows that the minimum and mean absolute percentage error between measured and calculated width are about 2.22% and 6.27%, respectively. The experiments and results are presented in this paper.Keywords: LiDAR, depth camera, real-time, detection and measurement
Procedia PDF Downloads 2322760 Application of Metric Dimension of Graph in Unraveling the Complexity of Hyperacusis
Authors: Hassan Ibrahim
Abstract:
The prevalence of hyperacusis, an auditory condition characterized by heightened sensitivity to sounds, continues to rise, posing challenges for effective diagnosis and intervention. It is believed that this work deepens will deepens the understanding of hyperacusis etiology by employing graph theory as a novel analytical framework. We constructed a comprehensive graph wherein nodes represent various factors associated with hyperacusis, including aging, head or neck trauma, infection/virus, depression, migraines, ear infection, anxiety, and other potential contributors. Relationships between factors are modeled as edges, allowing us to visualize and quantify the interactions within the etiological landscape of hyperacusis. it employ the concept of the metric dimension of a connected graph to identify key nodes (landmarks) that serve as critical influencers in the interconnected web of hyperacusis causes. This approach offers a unique perspective on the relative importance and centrality of different factors, shedding light on the complex interplay between physiological, psychological, and environmental determinants. Visualization techniques were also employed to enhance the interpretation and facilitate the identification of the central nodes. This research contributes to the growing body of knowledge surrounding hyperacusis by offering a network-centric perspective on its multifaceted causes. The outcomes hold the potential to inform clinical practices, guiding healthcare professionals in prioritizing interventions and personalized treatment plans based on the identified landmarks within the etiological network. Through the integration of graph theory into hyperacusis research, the complexity of this auditory condition was unraveled and pave the way for more effective approaches to its management.Keywords: auditory condition, connected graph, hyperacusis, metric dimension
Procedia PDF Downloads 402759 Evaluating the Satisfaction of Chinese Consumers toward Influencers at TikTok
Authors: Noriyuki Suyama
Abstract:
The progress and spread of digitalization have led to the provision of a variety of new services. The recent progress in digitization can be attributed to rapid developments in science and technology. First, the research and diffusion of artificial intelligence (AI) has made dramatic progress. Around 2000, the third wave of AI research, which had been underway for about 50 years, arrived. Specifically, machine learning and deep learning were made possible in AI, and the ability of AI to acquire knowledge, define the knowledge, and update its own knowledge in a quantitative manner made the use of big data practical even for commercial PCs. On the other hand, with the spread of social media, information exchange has become more common in our daily lives, and the lending and borrowing of goods and services, in other words, the sharing economy, has become widespread. The scope of this trend is not limited to any industry, and its momentum is growing as the SDGs take root. In addition, the Social Network Service (SNS), a part of social media, has brought about the evolution of the retail business. In the past few years, social network services (SNS) involving users or companies have especially flourished. The People's Republic of China (hereinafter referred to as "China") is a country that is stimulating enormous consumption through its own unique SNS, which is different from the SNS used in developed countries around the world. This paper focuses on the effectiveness and challenges of influencer marketing by focusing on the influence of influencers on users' behavior and satisfaction with Chinese SNSs. Specifically, Conducted was the quantitative survey of Tik Tok users living in China, with the aim of gaining new insights from the analysis and discussions. As a result, we found several important findings and knowledge.Keywords: customer satisfaction, social networking services, influencer marketing, Chinese consumers’ behavior
Procedia PDF Downloads 902758 Machine Learning Prediction of Compressive Damage and Energy Absorption in Carbon Fiber-Reinforced Polymer Tubular Structures
Authors: Milad Abbasi
Abstract:
Carbon fiber-reinforced polymer (CFRP) composite structures are increasingly being utilized in the automotive industry due to their lightweight and specific energy absorption capabilities. Although it is impossible to predict composite mechanical properties directly using theoretical methods, various research has been conducted so far in the literature for accurate simulation of CFRP structures' energy-absorbing behavior. In this research, axial compression experiments were carried out on hand lay-up unidirectional CFRP composite tubes. The fabrication method allowed the authors to extract the material properties of the CFRPs using ASTM D3039, D3410, and D3518 standards. A neural network machine learning algorithm was then utilized to build a robust prediction model to forecast the axial compressive properties of CFRP tubes while reducing high-cost experimental efforts. The predicted results have been compared with the experimental outcomes in terms of load-carrying capacity and energy absorption capability. The results showed high accuracy and precision in the prediction of the energy-absorption capacity of the CFRP tubes. This research also demonstrates the effectiveness and challenges of machine learning techniques in the robust simulation of composites' energy-absorption behavior. Interestingly, the proposed method considerably condensed numerical and experimental efforts in the simulation and calibration of CFRP composite tubes subjected to compressive loading.Keywords: CFRP composite tubes, energy absorption, crushing behavior, machine learning, neural network
Procedia PDF Downloads 1552757 Research Progress of the Relationship between Urban Rail Transit and Residents' Travel Behavior during 1999-2019: A Scientific Knowledge Mapping Based on Citespace and Vosviewer
Authors: Zheng Yi
Abstract:
Among the attempts made worldwide to foster urban and transport sustainability, transit-oriented development certainly is one of the most successful. Residents' travel behavior is a concern in the researches about the impacts of transit-oriented development. The study takes 620 English journal papers in the core collection database of Web of Science as the study objects; the paper tries to map out the scientific knowledge mapping in the field and draw the basic conditions by co-citation analysis, co-word analysis, a total of citation network analysis and visualization techniques. This study teases out the research hotspots and evolution of the relationship between urban rail transit and resident's travel behavior from 1999 to 2019. According to the results of the analysis of the time-zone view and burst-detection, the paper discusses the trend of the next stage of international study. The results show that in the past 20 years, the research focuses on these keywords: land use, behavior, model, built environment, impact, travel behavior, walking, physical activity, smart card, big data, simulation, perception. According to different research contents, the key literature is further divided into these topics: the attributes of the built environment, land use, transportation network, transportation policies. The results of this paper can help to understand the related researches and achievements systematically. These results can also provide a reference for identifying the main challenges that relevant researches need to address in the future.Keywords: urban rail transit, travel behavior, knowledge map, evolution of researches
Procedia PDF Downloads 1112756 Pavement Failures and Its Maintenance
Authors: Maulik L. Sisodia, Tirth K. Raval, Aarsh S. Mistry
Abstract:
This paper summarizes the ongoing researches about the defects in both flexible and rigid pavement and the maintenance in both flexible and rigid pavements. Various defects in pavements have been identified since the existence of both flexible and rigid pavement. Flexible Pavement failure is defined in terms of decreasing serviceability caused by the development of cracks, ruts, potholes etc. Flexible Pavement structure can be destroyed in a single season due to water penetration. Defects in flexible pavements is a problem of multiple dimensions, phenomenal growth of vehicular traffic (in terms of no. of axle loading of commercial vehicles), the rapid expansion in the road network, non-availability of suitable technology, material, equipment, skilled labor and poor funds allocation have all added complexities to the problem of flexible pavements. In rigid pavements due to different type of destress the failure like joint spalling, faulting, shrinkage cracking, punch out, corner break etc. Application of correction in the existing surface will enhance the life of maintenance works as well as that of strengthening layer. Maintenance of a road network involves a variety of operations, i.e., identification of deficiencies and planning, programming and scheduling for actual implementation in the field and monitoring. The essential objective should be to keep the road surface and appurtenances in good condition and to extend the life of the road assets to its design life. The paper describes lessons learnt from pavement failures and problems experienced during the last few years on a number of projects in India. Broadly, the activities include identification of defects and the possible cause there off, determination of appropriate remedial measures; implement these in the field and monitoring of the results.Keywords: Flexible Pavements, Rigid Pavements, Defects, Maintenance
Procedia PDF Downloads 1752755 Unraveling the Complexity of Hyperacusis: A Metric Dimension of a Graph Concept
Authors: Hassan Ibrahim
Abstract:
The prevalence of hyperacusis, an auditory condition characterized by heightened sensitivity to sounds, continues to rise, posing challenges for effective diagnosis and intervention. It is believed that this work deepens will deepens the understanding of hyperacusis etiology by employing graph theory as a novel analytical framework. it constructed a comprehensive graph wherein nodes represent various factors associated with hyperacusis, including aging, head or neck trauma, infection/virus, depression, migraines, ear infection, anxiety, and other potential contributors. Relationships between factors are modeled as edges, allowing us to visualize and quantify the interactions within the etiological landscape of hyperacusis. it employ the concept of the metric dimension of a connected graph to identify key nodes (landmarks) that serve as critical influencers in the interconnected web of hyperacusis causes. This approach offers a unique perspective on the relative importance and centrality of different factors, shedding light on the complex interplay between physiological, psychological, and environmental determinants. Visualization techniques were also employed to enhance the interpretation and facilitate the identification of the central nodes. This research contributes to the growing body of knowledge surrounding hyperacusis by offering a network-centric perspective on its multifaceted causes. The outcomes hold the potential to inform clinical practices, guiding healthcare professionals in prioritizing interventions and personalized treatment plans based on the identified landmarks within the etiological network. Through the integration of graph theory into hyperacusis research, the complexity of this auditory condition was unraveled and pave the way for more effective approaches to its management.Keywords: auditory condition, connected graph, hyperacusis, metric dimension
Procedia PDF Downloads 282754 Assessing Natura 2000 Network Effectiveness in Landscape Conservation: A Case Study in Castile and León, Spain (1990-2018)
Authors: Paula García-Llamas, Polonia Díez González, Angela Taboada
Abstract:
In an era marked by unprecedented anthropogenic alterations to landscapes and biodiversity, the consequential loss of fauna, flora, and habitats poses a grave concern. It is imperative to evaluate our capacity to manage and mitigate such changes effectively. This study aims to scrutinize the efficacy of the Natura 2000 Network (NN2000) in landscape conservation within the autonomous community of Castile and Leon (Spain), spanning from 1990 to 2018. Leveraging land use change maps from the European Corine Land Cover database across four subperiods (1990-2000, 2000-2006, 2006-2012, and 2012-2018), we quantified alterations occurring both within NN2000 protected sites and within a 5km buffer zone. Additionally, we spatially assess land use/land cover patterns of change considering fluxes of various habitat types defined within NN2000. Our findings reveal that the protected areas under NN2000 were particularly susceptible to change, with the most significant transformations observed during the 1990-2000 period. Predominant change processes include secondary succession and scrubland formation due to land use cessation, deforestation, and agricultural intensification. While NN2000 demonstrates efficacy in curtailing urbanization and industrialization within buffer zones, its management measures have proven insufficient in safeguarding landscapes against the dynamic changes witnessed between 1990 and 2018, especially in relation to rural abandonment.Keywords: Corine land cover, land cover changes, site of community importance, special protection area
Procedia PDF Downloads 502753 Mammographic Multi-View Cancer Identification Using Siamese Neural Networks
Authors: Alisher Ibragimov, Sofya Senotrusova, Aleksandra Beliaeva, Egor Ushakov, Yuri Markin
Abstract:
Mammography plays a critical role in screening for breast cancer in women, and artificial intelligence has enabled the automatic detection of diseases in medical images. Many of the current techniques used for mammogram analysis focus on a single view (mediolateral or craniocaudal view), while in clinical practice, radiologists consider multiple views of mammograms from both breasts to make a correct decision. Consequently, computer-aided diagnosis (CAD) systems could benefit from incorporating information gathered from multiple views. In this study, the introduce a method based on a Siamese neural network (SNN) model that simultaneously analyzes mammographic images from tri-view: bilateral and ipsilateral. In this way, when a decision is made on a single image of one breast, attention is also paid to two other images – a view of the same breast in a different projection and an image of the other breast as well. Consequently, the algorithm closely mimics the radiologist's practice of paying attention to the entire examination of a patient rather than to a single image. Additionally, to the best of our knowledge, this research represents the first experiments conducted using the recently released Vietnamese dataset of digital mammography (VinDr-Mammo). On an independent test set of images from this dataset, the best model achieved an AUC of 0.87 per image. Therefore, this suggests that there is a valuable automated second opinion in the interpretation of mammograms and breast cancer diagnosis, which in the future may help to alleviate the burden on radiologists and serve as an additional layer of verification.Keywords: breast cancer, computer-aided diagnosis, deep learning, multi-view mammogram, siamese neural network
Procedia PDF Downloads 1392752 Buffer Allocation and Traffic Shaping Policies Implemented in Routers Based on a New Adaptive Intelligent Multi Agent Approach
Authors: M. Taheri Tehrani, H. Ajorloo
Abstract:
In this paper, an intelligent multi-agent framework is developed for each router in which agents have two vital functionalities, traffic shaping and buffer allocation and are positioned in the ports of the routers. With traffic shaping functionality agents shape the traffic forward by dynamic and real time allocation of the rate of generation of tokens in a Token Bucket algorithm and with buffer allocation functionality agents share their buffer capacity between each other based on their need and the conditions of the network. This dynamic and intelligent framework gives this opportunity to some ports to work better under burst and more busy conditions. These agents work intelligently based on Reinforcement Learning (RL) algorithm and will consider effective parameters in their decision process. As RL have limitation considering much parameter in its decision process due to the volume of calculations, we utilize our novel method which invokes Principle Component Analysis (PCA) on the RL and gives a high dimensional ability to this algorithm to consider as much as needed parameters in its decision process. This implementation when is compared to our previous work where traffic shaping was done without any sharing and dynamic allocation of buffer size for each port, the lower packet drop in the whole network specifically in the source routers can be seen. These methods are implemented in our previous proposed intelligent simulation environment to be able to compare better the performance metrics. The results obtained from this simulation environment show an efficient and dynamic utilization of resources in terms of bandwidth and buffer capacities pre allocated to each port.Keywords: principal component analysis, reinforcement learning, buffer allocation, multi- agent systems
Procedia PDF Downloads 5192751 Comparing Deep Architectures for Selecting Optimal Machine Translation
Authors: Despoina Mouratidis, Katia Lida Kermanidis
Abstract:
Machine translation (MT) is a very important task in Natural Language Processing (NLP). MT evaluation is crucial in MT development, as it constitutes the means to assess the success of an MT system, and also helps improve its performance. Several methods have been proposed for the evaluation of (MT) systems. Some of the most popular ones in automatic MT evaluation are score-based, such as the BLEU score, and others are based on lexical similarity or syntactic similarity between the MT outputs and the reference involving higher-level information like part of speech tagging (POS). This paper presents a language-independent machine learning framework for classifying pairwise translations. This framework uses vector representations of two machine-produced translations, one from a statistical machine translation model (SMT) and one from a neural machine translation model (NMT). The vector representations consist of automatically extracted word embeddings and string-like language-independent features. These vector representations used as an input to a multi-layer neural network (NN) that models the similarity between each MT output and the reference, as well as between the two MT outputs. To evaluate the proposed approach, a professional translation and a "ground-truth" annotation are used. The parallel corpora used are English-Greek (EN-GR) and English-Italian (EN-IT), in the educational domain and of informal genres (video lecture subtitles, course forum text, etc.) that are difficult to be reliably translated. They have tested three basic deep learning (DL) architectures to this schema: (i) fully-connected dense, (ii) Convolutional Neural Network (CNN), and (iii) Long Short-Term Memory (LSTM). Experiments show that all tested architectures achieved better results when compared against those of some of the well-known basic approaches, such as Random Forest (RF) and Support Vector Machine (SVM). Better accuracy results are obtained when LSTM layers are used in our schema. In terms of a balance between the results, better accuracy results are obtained when dense layers are used. The reason for this is that the model correctly classifies more sentences of the minority class (SMT). For a more integrated analysis of the accuracy results, a qualitative linguistic analysis is carried out. In this context, problems have been identified about some figures of speech, as the metaphors, or about certain linguistic phenomena, such as per etymology: paronyms. It is quite interesting to find out why all the classifiers led to worse accuracy results in Italian as compared to Greek, taking into account that the linguistic features employed are language independent.Keywords: machine learning, machine translation evaluation, neural network architecture, pairwise classification
Procedia PDF Downloads 1332750 A Hybrid Artificial Intelligence and Two Dimensional Depth Averaged Numerical Model for Solving Shallow Water and Exner Equations Simultaneously
Authors: S. Mehrab Amiri, Nasser Talebbeydokhti
Abstract:
Modeling sediment transport processes by means of numerical approach often poses severe challenges. In this way, a number of techniques have been suggested to solve flow and sediment equations in decoupled, semi-coupled or fully coupled forms. Furthermore, in order to capture flow discontinuities, a number of techniques, like artificial viscosity and shock fitting, have been proposed for solving these equations which are mostly required careful calibration processes. In this research, a numerical scheme for solving shallow water and Exner equations in fully coupled form is presented. First-Order Centered scheme is applied for producing required numerical fluxes and the reconstruction process is carried out toward using Monotonic Upstream Scheme for Conservation Laws to achieve a high order scheme. In order to satisfy C-property of the scheme in presence of bed topography, Surface Gradient Method is proposed. Combining the presented scheme with fourth order Runge-Kutta algorithm for time integration yields a competent numerical scheme. In addition, to handle non-prismatic channels problems, Cartesian Cut Cell Method is employed. A trained Multi-Layer Perceptron Artificial Neural Network which is of Feed Forward Back Propagation (FFBP) type estimates sediment flow discharge in the model rather than usual empirical formulas. Hydrodynamic part of the model is tested for showing its capability in simulation of flow discontinuities, transcritical flows, wetting/drying conditions and non-prismatic channel flows. In this end, dam-break flow onto a locally non-prismatic converging-diverging channel with initially dry bed conditions is modeled. The morphodynamic part of the model is verified simulating dam break on a dry movable bed and bed level variations in an alluvial junction. The results show that the model is capable in capturing the flow discontinuities, solving wetting/drying problems even in non-prismatic channels and presenting proper results for movable bed situations. It can also be deducted that applying Artificial Neural Network, instead of common empirical formulas for estimating sediment flow discharge, leads to more accurate results.Keywords: artificial neural network, morphodynamic model, sediment continuity equation, shallow water equations
Procedia PDF Downloads 1902749 A Vision-Based Early Warning System to Prevent Elephant-Train Collisions
Authors: Shanaka Gunasekara, Maleen Jayasuriya, Nalin Harischandra, Lilantha Samaranayake, Gamini Dissanayake
Abstract:
One serious facet of the worsening Human-Elephant conflict (HEC) in nations such as Sri Lanka involves elephant-train collisions. Endangered Asian elephants are maimed or killed during such accidents, which also often result in orphaned or disabled elephants, contributing to the phenomenon of lone elephants. These lone elephants are found to be more likely to attack villages and showcase aggressive behaviour, which further exacerbates the overall HEC. Furthermore, Railway Services incur significant financial losses and disruptions to services annually due to such accidents. Most elephant-train collisions occur due to a lack of adequate reaction time. This is due to the significant stopping distance requirements of trains, as the full braking force needs to be avoided to minimise the risk of derailment. Thus, poor driver visibility at sharp turns, nighttime operation, and poor weather conditions are often contributing factors to this problem. Initial investigations also indicate that most collisions occur in localised “hotspots” where elephant pathways/corridors intersect with railway tracks that border grazing land and watering holes. Taking these factors into consideration, this work proposes the leveraging of recent developments in Convolutional Neural Network (CNN) technology to detect elephants using an RGB/infrared capable camera around known hotspots along the railway track. The CNN was trained using a curated dataset of elephants collected on field visits to elephant sanctuaries and wildlife parks in Sri Lanka. With this vision-based detection system at its core, a prototype unit of an early warning system was designed and tested. This weatherised and waterproofed unit consists of a Reolink security camera which provides a wide field of view and range, an Nvidia Jetson Xavier computing unit, a rechargeable battery, and a solar panel for self-sufficient functioning. The prototype unit was designed to be a low-cost, low-power and small footprint device that can be mounted on infrastructures such as poles or trees. If an elephant is detected, an early warning message is communicated to the train driver using the GSM network. A mobile app for this purpose was also designed to ensure that the warning is clearly communicated. A centralized control station manages and communicates all information through the train station network to ensure coordination among important stakeholders. Initial results indicate that detection accuracy is sufficient under varying lighting situations, provided comprehensive training datasets that represent a wide range of challenging conditions are available. The overall hardware prototype was shown to be robust and reliable. We envision a network of such units may help contribute to reducing the problem of elephant-train collisions and has the potential to act as an important surveillance mechanism in dealing with the broader issue of human-elephant conflicts.Keywords: computer vision, deep learning, human-elephant conflict, wildlife early warning technology
Procedia PDF Downloads 2272748 The Quality of Business Relationships in the Tourism System: An Imaginary Organisation Approach
Authors: Armando Luis Vieira, Carlos Costa, Arthur Araújo
Abstract:
The tourism system is viewable as a network of relationships amongst business partners where the success of each actor will ultimately be determined by the success of the whole network. Especially since the publication of Gümmesson’s (1996) ‘theory of imaginary organisations’, which suggests that organisational effectiveness largely depends on managing relationships and sharing resources and activities, relationship quality (RQ) has been increasingly recognised as a main source of value creation and competitive advantage. However, there is still ambiguity around this topic, and managers and researchers have been recurrently reporting the need to better understand and capitalise on the quality of interactions with business partners. This research aims at testing an RQ model from a relational, imaginary organisation’s approach. Two mail surveys provide the perceptions of 725 hotel representatives about their business relationships with tour operators, and 1,224 corporate client representatives about their business relationships with hotels (21.9 % and 38.8 % response rate, respectively). The analysis contributes to enhance our understanding on the linkages between RQ and its determinants, and identifies the role of their dimensions. Structural equation modelling results highlight trust as the dominant dimension, the crucial role of commitment and satisfaction, and suggest customer orientation as complementary building block. Findings also emphasise problem solving behaviour and selling orientation as the most relevant dimensions of customer orientation. The comparison of the two ‘dyads’ deepens the discussion and enriches the suggested theoretical and managerial guidelines concerning the contribution of quality relationships to business performance.Keywords: corporate clients, destination competitiveness, hotels, relationship quality, structural equations modelling, tour operators
Procedia PDF Downloads 3962747 Implementation of Social Network Analysis to Analyze the Dependency between Construction Bid Packages
Authors: Kawalpreet Kaur, Panagiotis Mitropoulos
Abstract:
The division of the project scope into work packages is the most important step in the preconstruction phase of construction projects. The work division determines the scope and complexity of each bid package, resulting in dependencies between project participants performing these work packages. The coordination between project participants is necessary because of these dependencies. Excessive dependencies between the bid packages create coordination difficulties, leading to delays, added costs, and contractual friction among project participants. However, the literature on construction provides limited knowledge regarding work structuring approaches, issues, and challenges. Manufacturing industry literature provides a systematic approach to defining the project scope into work packages, and the implementation of social network analysis (SNA) in manufacturing is an effective approach to defining and analyzing the divided scope of work at the dependencies level. This paper presents a case study of implementing a similar approach using SNA in construction bid packages. The study uses SNA to analyze the scope of bid packages and determine the dependency between scope elements. The method successfully identifies the bid package with the maximum interaction with other trade contractors and the scope elements that are crucial for project performance. The analysis provided graphical and quantitative information on bid package dependencies. The study can be helpful in performing an analysis to determine the dependencies between bid packages and their scope elements and how these scope elements are critical for project performance. The study illustrates the potential use of SNA as a systematic approach to analyzing bid package dependencies in construction projects, which can guide the division of crucial scope elements to minimize negative impacts on project performance.Keywords: work structuring, bid packages, work breakdown, project participants
Procedia PDF Downloads 812746 Computational Insights Into Allosteric Regulation of Lyn Protein Kinase: Structural Dynamics and Impacts of Cancer-Related Mutations
Authors: Mina Rabipour, Elena Pallaske, Floyd Hassenrück, Rocio Rebollido-Rios
Abstract:
Protein tyrosine kinases, including Lyn kinase of the Src family kinases (SFK), regulate cell proliferation, survival, and differentiation. Lyn kinase has been implicated in various cancers, positioning it as a promising therapeutic target. However, the conserved ATP-binding pocket across SFKs makes developing selective inhibitors challenging. This study aims to address this limitation by exploring the potential for allosteric modulation of Lyn kinase, focusing on how its structural dynamics and specific oncogenic mutations impact its conformation and function. To achieve this, we combined homology modeling, molecular dynamics simulations, and data science techniques to conduct microsecond-length simulations. Our approach allowed a detailed investigation into the interplay between Lyn’s catalytic and regulatory domains, identifying key conformational states involved in allosteric regulation. Additionally, we evaluated the structural effects of Dasatinib, a competitive inhibitor, and ATP binding on Lyn active conformation. Notably, our simulations show that cancer-related mutations, specifically I364L/N and E290D/K, shift Lyn toward an inactive conformation, contrasting with the active state of the wild-type protein. This may suggest how these mutations contribute to aberrant signaling in cancer cells. We conducted a dynamical network analysis to assess residue-residue interactions and the impact of mutations on the Lyn intramolecular network. This revealed significant disruptions due to mutations, especially in regions distant from the ATP-binding site. These disruptions suggest potential allosteric sites as therapeutic targets, offering an alternative strategy for Lyn inhibition with higher specificity and fewer off-target effects compared to ATP-competitive inhibitors. Our findings provide insights into Lyn kinase regulation and highlight allosteric sites as avenues for selective drug development. Targeting these sites may modulate Lyn activity in cancer cells, reducing toxicity and improving outcomes. Furthermore, our computational strategy offers a scalable approach for analyzing other SFK members or kinases with similar properties, facilitating the discovery of selective allosteric modulators and contributing to precise cancer therapies.Keywords: lyn tyrosine kinase, mutation analysis, conformational changes, dynamic network analysis, allosteric modulation, targeted inhibition
Procedia PDF Downloads 172745 Quantified Metabolomics for the Determination of Phenotypes and Biomarkers across Species in Health and Disease
Authors: Miroslava Cuperlovic-Culf, Lipu Wang, Ketty Boyle, Nadine Makley, Ian Burton, Anissa Belkaid, Mohamed Touaibia, Marc E. Surrette
Abstract:
Metabolic changes are one of the major factors in the development of a variety of diseases in various species. Metabolism of agricultural plants is altered the following infection with pathogens sometimes contributing to resistance. At the same time, pathogens use metabolites for infection and progression. In humans, metabolism is a hallmark of cancer development for example. Quantified metabolomics data combined with other omics or clinical data and analyzed using various unsupervised and supervised methods can lead to better diagnosis and prognosis. It can also provide information about resistance as well as contribute knowledge of compounds significant for disease progression or prevention. In this work, different methods for metabolomics quantification and analysis from Nuclear Magnetic Resonance (NMR) measurements that are used for investigation of disease development in wheat and human cells will be presented. One-dimensional 1H NMR spectra are used extensively for metabolic profiling due to their high reliability, wide range of applicability, speed, trivial sample preparation and low cost. This presentation will describe a new method for metabolite quantification from NMR data that combines alignment of spectra of standards to sample spectra followed by multivariate linear regression optimization of spectra of assigned metabolites to samples’ spectra. Several different alignment methods were tested and multivariate linear regression result has been compared with other quantification methods. Quantified metabolomics data can be analyzed in the variety of ways and we will present different clustering methods used for phenotype determination, network analysis providing knowledge about the relationships between metabolites through metabolic network as well as biomarker selection providing novel markers. These analysis methods have been utilized for the investigation of fusarium head blight resistance in wheat cultivars as well as analysis of the effect of estrogen receptor and carbonic anhydrase activation and inhibition on breast cancer cell metabolism. Metabolic changes in spikelet’s of wheat cultivars FL62R1, Stettler, MuchMore and Sumai3 following fusarium graminearum infection were explored. Extensive 1D 1H and 2D NMR measurements provided information for detailed metabolite assignment and quantification leading to possible metabolic markers discriminating resistance level in wheat subtypes. Quantification data is compared to results obtained using other published methods. Fusarium infection induced metabolic changes in different wheat varieties are discussed in the context of metabolic network and resistance. Quantitative metabolomics has been used for the investigation of the effect of targeted enzyme inhibition in cancer. In this work, the effect of 17 β -estradiol and ferulic acid on metabolism of ER+ breast cancer cells has been compared to their effect on ER- control cells. The effect of the inhibitors of carbonic anhydrase on the observed metabolic changes resulting from ER activation has also been determined. Metabolic profiles were studied using 1D and 2D metabolomic NMR experiments, combined with the identification and quantification of metabolites, and the annotation of the results is provided in the context of biochemical pathways.Keywords: metabolic biomarkers, metabolic network, metabolomics, multivariate linear regression, NMR quantification, quantified metabolomics, spectral alignment
Procedia PDF Downloads 3402744 Potentiality of the Wind Energy in Algeria
Authors: C. Benoudjafer, M. N. Tandjaoui, C. Benachaiba
Abstract:
The use of kinetic energy of the wind is in full rise in the world and it starts to be known in our country but timidly. One or more aero generators can be installed to produce for example electricity on isolated places or not connected to the electrical supply network. To use the wind as energy source, it is necessary to know first the energy needs for the population and study the wind intensity, speed, frequency and direction.Keywords: Algeria, renewable energies, wind, wind power, aero-generators, wind energetic potential
Procedia PDF Downloads 4332743 Structure Clustering for Milestoning Applications of Complex Conformational Transitions
Authors: Amani Tahat, Serdal Kirmizialtin
Abstract:
Trajectory fragment methods such as Markov State Models (MSM), Milestoning (MS) and Transition Path sampling are the prime choice of extending the timescale of all atom Molecular Dynamics simulations. In these approaches, a set of structures that covers the accessible phase space has to be chosen a priori using cluster analysis. Structural clustering serves to partition the conformational state into natural subgroups based on their similarity, an essential statistical methodology that is used for analyzing numerous sets of empirical data produced by Molecular Dynamics (MD) simulations. Local transition kernel among these clusters later used to connect the metastable states using a Markovian kinetic model in MSM and a non-Markovian model in MS. The choice of clustering approach in constructing such kernel is crucial since the high dimensionality of the biomolecular structures might easily confuse the identification of clusters when using the traditional hierarchical clustering methodology. Of particular interest, in the case of MS where the milestones are very close to each other, accurate determination of the milestone identity of the trajectory becomes a challenging issue. Throughout this work we present two cluster analysis methods applied to the cis–trans isomerism of dinucleotide AA. The choice of nucleic acids to commonly used proteins to study the cluster analysis is two fold: i) the energy landscape is rugged; hence transitions are more complex, enabling a more realistic model to study conformational transitions, ii) Nucleic acids conformational space is high dimensional. A diverse set of internal coordinates is necessary to describe the metastable states in nucleic acids, posing a challenge in studying the conformational transitions. Herein, we need improved clustering methods that accurately identify the AA structure in its metastable states in a robust way for a wide range of confused data conditions. The single linkage approach of the hierarchical clustering available in GROMACS MD-package is the first clustering methodology applied to our data. Self Organizing Map (SOM) neural network, that also known as a Kohonen network, is the second data clustering methodology. The performance comparison of the neural network as well as hierarchical clustering method is studied by means of computing the mean first passage times for the cis-trans conformational rates. Our hope is that this study provides insight into the complexities and need in determining the appropriate clustering algorithm for kinetic analysis. Our results can improve the effectiveness of decisions based on clustering confused empirical data in studying conformational transitions in biomolecules.Keywords: milestoning, self organizing map, single linkage, structure clustering
Procedia PDF Downloads 2242742 Accessibility Assessment of School Facilities Using Geospatial Technologies: A Case Study of District Sheikhupura
Authors: Hira Jabbar
Abstract:
Education is vital for inclusive growth of an economy and a critical contributor for investment in human capital. Like other developing countries, Pakistan is facing enormous challenges regarding the provision of public facilities, improper infrastructure planning, accelerating rate of population and poor accessibility. The influence of the rapid advancement and innovations in GIS and RS techniques have proved to be a useful tool for better planning and decision making to encounter these challenges. Therefore present study incorporates GIS and RS techniques to investigate the spatial distribution of school facilities, identifies settlements with served and unserved population, finds potential areas for new schools based on population and develops an accessibility index to evaluate the higher accessibility for schools. For this purpose high-resolution worldview imagery was used to develop road network, settlements and school facilities and to generate school accessibility for each level. Landsat 8 imagery was utilized to extract built-up area by applying pre and post-processing models and Landscan 2015 was used to analyze population statistics. Service area analysis was performed using network analyst extension in ArcGIS 10.3v and results were evaluated for served and underserved areas and population. An accessibility tool was used to evaluate a set of potential destinations to determine which is the most accessible with the given population distribution. Findings of the study may contribute to facilitating the town planners and education authorities for understanding the existing patterns of school facilities. It is concluded that GIS and remote sensing can be effectively used in urban transport and facility planning.Keywords: accessibility, geographic information system, landscan, worldview
Procedia PDF Downloads 3272741 Prediction of California Bearing Ratio of a Black Cotton Soil Stabilized with Waste Glass and Eggshell Powder using Artificial Neural Network
Authors: Biruhi Tesfaye, Avinash M. Potdar
Abstract:
The laboratory test process to determine the California bearing ratio (CBR) of black cotton soils is not only overpriced but also time-consuming as well. Hence advanced prediction of CBR plays a significant role as it is applicable In pavement design. The prediction of CBR of treated soil was executed by Artificial Neural Networks (ANNs) which is a Computational tool based on the properties of the biological neural system. To observe CBR values, combined eggshell and waste glass was added to soil as 4, 8, 12, and 16 % of the weights of the soil samples. Accordingly, the laboratory related tests were conducted to get the required best model. The maximum CBR value found at 5.8 at 8 % of eggshell waste glass powder addition. The model was developed using CBR as an output layer variable. CBR was considered as a function of the joint effect of liquid limit, plastic limit, and plastic index, optimum moisture content and maximum dry density. The best model that has been found was ANN with 5, 6 and 1 neurons in the input, hidden and output layer correspondingly. The performance of selected ANN has been 0.99996, 4.44E-05, 0.00353 and 0.0067 which are correlation coefficient (R), mean square error (MSE), mean absolute error (MAE) and root mean square error (RMSE) respectively. The research presented or summarized above throws light on future scope on stabilization with waste glass combined with different percentages of eggshell that leads to the economical design of CBR acceptable to pavement sub-base or base, as desired.Keywords: CBR, artificial neural network, liquid limit, plastic limit, maximum dry density, OMC
Procedia PDF Downloads 1932740 Maximum Induced Subgraph of an Augmented Cube
Authors: Meng-Jou Chien, Jheng-Cheng Chen, Chang-Hsiung Tsai
Abstract:
Let maxζG(m) denote the maximum number of edges in a subgraph of graph G induced by m nodes. The n-dimensional augmented cube, denoted as AQn, a variation of the hypercube, possesses some properties superior to those of the hypercube. We study the cases when G is the augmented cube AQn.Keywords: interconnection network, augmented cube, induced subgraph, bisection width
Procedia PDF Downloads 4072739 Multi-Temporal Mapping of Built-up Areas Using Daytime and Nighttime Satellite Images Based on Google Earth Engine Platform
Authors: S. Hutasavi, D. Chen
Abstract:
The built-up area is a significant proxy to measure regional economic growth and reflects the Gross Provincial Product (GPP). However, an up-to-date and reliable database of built-up areas is not always available, especially in developing countries. The cloud-based geospatial analysis platform such as Google Earth Engine (GEE) provides an opportunity with accessibility and computational power for those countries to generate the built-up data. Therefore, this study aims to extract the built-up areas in Eastern Economic Corridor (EEC), Thailand using day and nighttime satellite imagery based on GEE facilities. The normalized indices were generated from Landsat 8 surface reflectance dataset, including Normalized Difference Built-up Index (NDBI), Built-up Index (BUI), and Modified Built-up Index (MBUI). These indices were applied to identify built-up areas in EEC. The result shows that MBUI performs better than BUI and NDBI, with the highest accuracy of 0.85 and Kappa of 0.82. Moreover, the overall accuracy of classification was improved from 79% to 90%, and error of total built-up area was decreased from 29% to 0.7%, after night-time light data from the Visible and Infrared Imaging Suite (VIIRS) Day Night Band (DNB). The results suggest that MBUI with night-time light imagery is appropriate for built-up area extraction and be utilize for further study of socioeconomic impacts of regional development policy over the EEC region.Keywords: built-up area extraction, google earth engine, adaptive thresholding method, rapid mapping
Procedia PDF Downloads 1272738 Optimum Design for Cathode Microstructure of Solid Oxide Fuel Cell
Authors: M. Riazat, H. Abdolvand, M. Baniassadi
Abstract:
In this present work, 3D reconstruction of cathode of SOFC is developed with various volume fractions and porosity. Three Phase Boundary (TPB) of construction of such derived micro structures is calculated. The neural network is used to optimize the porosity and volume fraction of each phase to reach a structure with maximum TPB.Keywords: fuel cell, solid oxide, TPB, 3D reconstruction
Procedia PDF Downloads 3252737 Time and Cost Prediction Models for Language Classification Over a Large Corpus on Spark
Authors: Jairson Barbosa Rodrigues, Paulo Romero Martins Maciel, Germano Crispim Vasconcelos
Abstract:
This paper presents an investigation of the performance impacts regarding the variation of five factors (input data size, node number, cores, memory, and disks) when applying a distributed implementation of Naïve Bayes for text classification of a large Corpus on the Spark big data processing framework. Problem: The algorithm's performance depends on multiple factors, and knowing before-hand the effects of each factor becomes especially critical as hardware is priced by time slice in cloud environments. Objectives: To explain the functional relationship between factors and performance and to develop linear predictor models for time and cost. Methods: the solid statistical principles of Design of Experiments (DoE), particularly the randomized two-level fractional factorial design with replications. This research involved 48 real clusters with different hardware arrangements. The metrics were analyzed using linear models for screening, ranking, and measurement of each factor's impact. Results: Our findings include prediction models and show some non-intuitive results about the small influence of cores and the neutrality of memory and disks on total execution time, and the non-significant impact of data input scale on costs, although notably impacts the execution time.Keywords: big data, design of experiments, distributed machine learning, natural language processing, spark
Procedia PDF Downloads 1212736 Application of Continuum Damage Concept to Simulation of the Interaction between Hydraulic Fractures and Natural Fractures
Authors: Anny Zambrano, German Gonzalez, Yair Quintero
Abstract:
The continuum damage concept is used to study the interaction between hydraulic fractures and natural fractures, the objective is representing the path and relation among this two fractures types and predict its complex behavior without the need to pre-define their direction as occurs in other finite element applications, providing results more consistent with the physical behavior of the phenomenon. The approach uses finite element simulations through Abaqus software to model damage fracturing, the fracturing process by damage propagation in a rock. The modeling the phenomenon develops in two dimensional (2D) so that the fracture will be represented by a line and the crack front by a point. It considers nonlinear constitutive behavior, finite strain, time-dependent deformation, complex boundary conditions, strain hardening and softening, and strain based damage evolution in compression and tension. The complete governing equations are provided and the method is described in detail to permit readers to replicate all results. The model is compared to models that are published and available. Comparisons are focused in five interactions between natural fractures (NF) and hydraulic fractures: Fractured arrested at NF, crossing NF with or without offset, branching at intersecting NFs, branching at end of NF and NF dilation due to shear slippage. The most significant new finding is, that is not necessary to use pre-defined addresses propagation and stress condition can be evaluated as a dominant factor in the process. This is important because it can model in a more real way the generated complex hydraulic fractures, and be a valuable tool to predict potential problems and different geometries of the fracture network in the process of fracturing due to fluid injection.Keywords: continuum damage, hydraulic fractures, natural fractures, complex fracture network, stiffness
Procedia PDF Downloads 3442735 Green Supply Chain Network Optimization with Internet of Things
Authors: Sema Kayapinar, Ismail Karaoglan, Turan Paksoy, Hadi Gokcen
Abstract:
Green Supply Chain Management is gaining growing interest among researchers and supply chain management. The concept of Green Supply Chain Management is to integrate environmental thinking into the Supply Chain Management. It is the systematic concept emphasis on environmental problems such as reduction of greenhouse gas emissions, energy efficiency, recycling end of life products, generation of solid and hazardous waste. This study is to present a green supply chain network model integrated Internet of Things applications. Internet of Things provides to get precise and accurate information of end-of-life product with sensors and systems devices. The forward direction consists of suppliers, plants, distributions centres and sales and collect centres while, the reverse flow includes the sales and collects centres, disassembled centre, recycling and disposal centre. The sales and collection centre sells the new products are transhipped from factory via distribution centre and also receive the end-of life product according their value level. We describe green logistics activities by presenting specific examples including “recycling of the returned products and “reduction of CO2 gas emissions”. The different transportation choices are illustrated between echelons according to their CO2 gas emissions. This problem is formulated as a mixed integer linear programming model to solve the green supply chain problems which are emerged from the environmental awareness and responsibilities. This model is solved by using Gams package program. Numerical examples are suggested to illustrate the efficiency of the proposed model.Keywords: green supply chain optimization, internet of things, greenhouse gas emission, recycling
Procedia PDF Downloads 3302734 Covid Medical Imaging Trial: Utilising Artificial Intelligence to Identify Changes on Chest X-Ray of COVID
Authors: Leonard Tiong, Sonit Singh, Kevin Ho Shon, Sarah Lewis
Abstract:
Investigation into the use of artificial intelligence in radiology continues to develop at a rapid rate. During the coronavirus pandemic, the combination of an exponential increase in chest x-rays and unpredictable staff shortages resulted in a huge strain on the department's workload. There is a World Health Organisation estimate that two-thirds of the global population does not have access to diagnostic radiology. Therefore, there could be demand for a program that could detect acute changes in imaging compatible with infection to assist with screening. We generated a conventional neural network and tested its efficacy in recognizing changes compatible with coronavirus infection. Following ethics approval, a deidentified set of 77 normal and 77 abnormal chest x-rays in patients with confirmed coronavirus infection were used to generate an algorithm that could train, validate and then test itself. DICOM and PNG image formats were selected due to their lossless file format. The model was trained with 100 images (50 positive, 50 negative), validated against 28 samples (14 positive, 14 negative), and tested against 26 samples (13 positive, 13 negative). The initial training of the model involved training a conventional neural network in what constituted a normal study and changes on the x-rays compatible with coronavirus infection. The weightings were then modified, and the model was executed again. The training samples were in batch sizes of 8 and underwent 25 epochs of training. The results trended towards an 85.71% true positive/true negative detection rate and an area under the curve trending towards 0.95, indicating approximately 95% accuracy in detecting changes on chest X-rays compatible with coronavirus infection. Study limitations include access to only a small dataset and no specificity in the diagnosis. Following a discussion with our programmer, there are areas where modifications in the weighting of the algorithm can be made in order to improve the detection rates. Given the high detection rate of the program, and the potential ease of implementation, this would be effective in assisting staff that is not trained in radiology in detecting otherwise subtle changes that might not be appreciated on imaging. Limitations include the lack of a differential diagnosis and application of the appropriate clinical history, although this may be less of a problem in day-to-day clinical practice. It is nonetheless our belief that implementing this program and widening its scope to detecting multiple pathologies such as lung masses will greatly assist both the radiology department and our colleagues in increasing workflow and detection rate.Keywords: artificial intelligence, COVID, neural network, machine learning
Procedia PDF Downloads 952733 Mitigating the Negative Health Effects from Stress - A Social Network Analysis
Authors: Jennifer A. Kowalkowski
Abstract:
Production agriculture (farming) is a physically, emotionally, and cognitively stressful occupation, where workers have little control over the stressors that impact both their work and their lives. In an occupation already rife with hazards, these occupational-related stressors have been shown to increase farm workers’ risks for illness, injury, disability, and death associated with their work. Despite efforts to mitigate the negative health effects from occupational-related stress (ORS) and to promote health and well-being (HWB) among farmers in the US, marked improvements have not been attained. Social support accessed through social networks has been shown to buffer against the negative health effects from stress, yet no studies have directly examined these relationships among farmers. The purpose of this study was to use social network analysis to explore the social networks of farm owner-operators and the social supports available to them for mitigating the negative health effects of ORS. A convenience sample of 71 farm owner-operators from a Midwestern County in the US completed and returned a mailed survey (55.5% response rate) that solicited information about their social networks related to ORS. Farmers reported an average of 2.4 individuals in their personal networks and higher levels of comfort discussing ORS with female network members. Farmers also identified few connections (3.4% density) and indicated low comfort with members of affiliation networks specific to ORS. Findings from this study highlighted that farmers accessed different social networks and resources for their personal HWB than for issues related to occupational(farm-related) health and safety. In addition, farmers’ social networks for personal HWB were smaller, with different relational characteristics than reported in studies of farmers’ social networks related to occupational health and safety. Collectively, these findings suggest that farmers conceptualize personal HWB differently than farm health and safety. Therefore, the same research approaches and targets that guide occupational health and safety research may not be appropriate for personal HWB for farmers. Interventions and programming targeting ORS and HWB have largely been offered through the same platforms or mechanisms as occupational health and safety programs. This may be attributed to the significant overlap between the farm as a family business and place of residence, or that ORS stems from farm-related issues. However, these assumptions translated to health research of farmers and farm families from the occupational health and safety literature have not been directly studied or challenged. Thismay explain why past interventions have not been effective at improving health outcomes for farmers and farm families. A close examination of findings from this study raises important questions for researchers who study agricultural health. Findings from this study have significant implications for future research agendas focused on addressing ORS, HWB, and health disparities for farmersand farm families.Keywords: agricultural health, occupational-related stress, social networks, well-being
Procedia PDF Downloads 1102732 Application of Federated Learning in the Health Care Sector for Malware Detection and Mitigation Using Software-Defined Networking Approach
Authors: A. Dinelka Panagoda, Bathiya Bandara, Chamod Wijetunga, Chathura Malinda, Lakmal Rupasinghe, Chethana Liyanapathirana
Abstract:
This research takes us forward with the concepts of Federated Learning and Software-Defined Networking (SDN) to introduce an efficient malware detection technique and provide a mitigation mechanism to give birth to a resilient and automated healthcare sector network system by also adding the feature of extended privacy preservation. Due to the daily transformation of new malware attacks on hospital Integrated Clinical Environment (ICEs), the healthcare industry is at an undefinable peak of never knowing its continuity direction. The state of blindness by the array of indispensable opportunities that new medical device inventions and their connected coordination offer daily, a factor that should be focused driven is not yet entirely understood by most healthcare operators and patients. This solution has the involvement of four clients in the form of hospital networks to build up the federated learning experimentation architectural structure with different geographical participation to reach the most reasonable accuracy rate with privacy preservation. While the logistic regression with cross-entropy conveys the detection, SDN comes in handy in the second half of the research to stack up the initial development phases of the system with malware mitigation based on policy implementation. The overall evaluation sums up with a system that proves the accuracy with the added privacy. It is no longer needed to continue with traditional centralized systems that offer almost everything but not privacy.Keywords: software-defined network, federated learning, privacy, integrated clinical environment, decentralized learning, malware detection, malware mitigation
Procedia PDF Downloads 190