Search results for: co-authorship network analysis
27870 Storm-Runoff Simulation Approaches for External Natural Catchments of Urban Sewer Systems
Authors: Joachim F. Sartor
Abstract:
According to German guidelines, external natural catchments are greater sub-catchments without significant portions of impervious areas, which possess a surface drainage system and empty in a sewer network. Basically, such catchments should be disconnected from sewer networks, particularly from combined systems. If this is not possible due to local conditions, their flow hydrographs have to be considered at the design of sewer systems, because the impact may be significant. Since there is a lack of sufficient measurements of storm-runoff events for such catchments and hence verified simulation methods to analyze their design flows, German standards give only general advices and demands special considerations in such cases. Compared to urban sub-catchments, external natural catchments exhibit greatly different flow characteristics. With increasing area size their hydrological behavior approximates that of rural catchments, e.g. sub-surface flow may prevail and lag times are comparable long. There are few observed peak flow values and simple (mostly empirical) approaches that are offered by literature for Central Europe. Most of them are at least helpful to crosscheck results that are achieved by simulation lacking calibration. Using storm-runoff data from five monitored rural watersheds in the west of Germany with catchment areas between 0.33 and 1.07 km2 , the author investigated by multiple event simulation three different approaches to determine the rainfall excess. These are the modified SCS variable run-off coefficient methods by Lutz and Zaiß as well as the soil moisture model by Ostrowski. Selection criteria for storm events from continuous precipitation data were taken from recommendations of M 165 and the runoff concentration method (parallel cascades of linear reservoirs) from a DWA working report to which the author had contributed. In general, the two run-off coefficient methods showed results that are of sufficient accuracy for most practical purposes. The soil moisture model showed no significant better results, at least not to such a degree that it would justify the additional data collection that its parameter determination requires. Particularly typical convective summer events after long dry periods, that are often decisive for sewer networks (not so much for rivers), showed discrepancies between simulated and measured flow hydrographs.Keywords: external natural catchments, sewer network design, storm-runoff modelling, urban drainage
Procedia PDF Downloads 15327869 DenseNet and Autoencoder Architecture for COVID-19 Chest X-Ray Image Classification and Improved U-Net Lung X-Ray Segmentation
Authors: Jonathan Gong
Abstract:
Purpose AI-driven solutions are at the forefront of many pathology and medical imaging methods. Using algorithms designed to better the experience of medical professionals within their respective fields, the efficiency and accuracy of diagnosis can improve. In particular, X-rays are a fast and relatively inexpensive test that can diagnose diseases. In recent years, X-rays have not been widely used to detect and diagnose COVID-19. The under use of Xrays is mainly due to the low diagnostic accuracy and confounding with pneumonia, another respiratory disease. However, research in this field has expressed a possibility that artificial neural networks can successfully diagnose COVID-19 with high accuracy. Models and Data The dataset used is the COVID-19 Radiography Database. This dataset includes images and masks of chest X-rays under the labels of COVID-19, normal, and pneumonia. The classification model developed uses an autoencoder and a pre-trained convolutional neural network (DenseNet201) to provide transfer learning to the model. The model then uses a deep neural network to finalize the feature extraction and predict the diagnosis for the input image. This model was trained on 4035 images and validated on 807 separate images from the ones used for training. The images used to train the classification model include an important feature: the pictures are cropped beforehand to eliminate distractions when training the model. The image segmentation model uses an improved U-Net architecture. This model is used to extract the lung mask from the chest X-ray image. The model is trained on 8577 images and validated on a validation split of 20%. These models are calculated using the external dataset for validation. The models’ accuracy, precision, recall, f1-score, IOU, and loss are calculated. Results The classification model achieved an accuracy of 97.65% and a loss of 0.1234 when differentiating COVID19-infected, pneumonia-infected, and normal lung X-rays. The segmentation model achieved an accuracy of 97.31% and an IOU of 0.928. Conclusion The models proposed can detect COVID-19, pneumonia, and normal lungs with high accuracy and derive the lung mask from a chest X-ray with similarly high accuracy. The hope is for these models to elevate the experience of medical professionals and provide insight into the future of the methods used.Keywords: artificial intelligence, convolutional neural networks, deep learning, image processing, machine learning
Procedia PDF Downloads 13127868 Urine Neutrophil Gelatinase-Associated Lipocalin as an Early Marker of Acute Kidney Injury in Hematopoietic Stem Cell Transplantation Patients
Authors: Sara Ataei, Maryam Taghizadeh-Ghehi, Amir Sarayani, Asieh Ashouri, Amirhossein Moslehi, Molouk Hadjibabaie, Kheirollah Gholami
Abstract:
Background: Acute kidney injury (AKI) is common in hematopoietic stem cell transplantation (HSCT) patients with an incidence of 21–73%. Prevention and early diagnosis reduces the frequency and severity of this complication. Predictive biomarkers are of major importance to timely diagnosis. Neutrophil gelatinase associated lipocalin (NGAL) is a widely investigated novel biomarker for early diagnosis of AKI. However, no study assessed NGAL for AKI diagnosis in HSCT patients. Methods: We performed further analyses on gathered data from our recent trial to evaluate the performance of urine NGAL (uNGAL) as an indicator of AKI in 72 allogeneic HSCT patients. AKI diagnosis and severity were assessed using Risk–Injury–Failure–Loss–End-stage renal disease and AKI Network criteria. We assessed uNGAL on days -6, -3, +3, +9 and +15. Results: Time-dependent Cox regression analysis revealed a statistically significant relationship between uNGAL and AKI occurrence. (HR=1.04 (1.008-1.07), P=0.01). There was a relation between uNGAL day +9 to baseline ratio and incidence of AKI (unadjusted HR=.1.047(1.012-1.083), P<0.01). The area under the receiver-operating characteristic curve for day +9 to baseline ratio was 0.86 (0.74-0.99, P<0.01) and a cut-off value of 2.62 was 85% sensitive and 83% specific in predicting AKI. Conclusions: Our results indicated that increase in uNGAL augmented the risk of AKI and the changes of day +9 uNGAL concentrations from baseline could be of value for predicting AKI in HSCT patients. Additionally uNGAL changes preceded serum creatinine rises by nearly 2 days.Keywords: acute kidney injury, hemtopoietic stem cell transplantation, neutrophil gelatinase-associated lipocalin, Receiver-operating characteristic curve
Procedia PDF Downloads 41027867 Disowning of ‘Our Lady of Alice Bhatti’ by Mohammad Hanif Through Gendered and Religious Discourse
Authors: Abrar Ajmal
Abstract:
The language used in literature reveals the culture and social gestalt of any society in which it has been constructed and consumed. This paper carries the same rationale, which aims to track certain socio-religious and cultural-economic disparities and discrepancies towards minorities, particularly Christians, in an Islamic re(public) where there is a clear majority of Muslims with the help of analysis of instances of language used in the narratives “Our Lady of Alice Bhatt” by Mohammad Hanif. It would highlight social inequalities practiced deeply in sociocultural discourse. Moreover, this research would also touch upon the question of gender discrimination and gender construction as a female entity in a male-chauvinistic scenic turnout using language since the novel revolves around communicative forfeits of Alice Bhatti’s life where she is fraying in fisticuffs to befit herself in a miss-fitted society. It would employ using Fairclough's framework for analysis to conduct a critical discourse analysis of the text at three axiom levels namely textual analysis, discursive practices, and socio-cultural analysis. Thus, the results would reveal textual findings in linguistic analysis, a range of embedded discourses in discursive practices, and consumption of the text into socio-cultural explications with the use of language and lexicalization employed in the selected excerpts.Keywords: gendered discourse, socio-economic disparities minorities, Islamization, analytical framework
Procedia PDF Downloads 6127866 Modern Trends in Foreign Direct Investments in Georgia
Authors: Rusudan Kinkladze, Guguli Kurashvili, Ketevan Chitaladze
Abstract:
Foreign direct investment is a driving force in the development of the interdependent national economies, and the study and analysis of investments is an urgent problem. It is particularly important for transitional economies, such as Georgia, and the study and analysis of investments is an urgent problem. Consequently, the goal of the research is the study and analysis of direct foreign investments in Georgia, and identification and forecasting of modern trends, and covers the period of 2006-2015. The study uses the methods of statistical observation, grouping and analysis, the methods of analytical indicators of time series, trend identification and the predicted values are calculated, as well as various literary and Internet sources relevant to the research. The findings showed that modern investment policy In Georgia is favorable for domestic as well as foreign investors. Georgia is still a net importer of investments. In 2015, the top 10 investing countries was led by Azerbaijan, United Kingdom and Netherlands, and the largest share of FDIs were allocated in the transport and communication sector; the financial sector was the second, followed by the health and social work sector, and the same trend will continue in the future.Keywords: foreign direct investments, methods, statistics, analysis
Procedia PDF Downloads 33327865 The Use of Multivariate Statistical and GIS for Characterization Groundwater Quality in Laghouat Region, Algeria
Authors: Rouighi Mustapha, Bouzid Laghaa Souad, Rouighi Tahar
Abstract:
Due to rain Shortage and the increase of population in the last years, wells excavation and groundwater use for different purposes had been increased without any planning. This is a great challenge for our country. Moreover, this scarcity of water resources in this region is unfortunately combined with rapid fresh water resources quality deterioration, due to salinity and contamination processes. Therefore, it is necessary to conduct the studies about groundwater quality in Algeria. In this work consists in the identification of the factors which influence the water quality parameters in Laghouat region by using statistical analysis Principal Component Analysis (PCA), Hierarchical Cluster Analysis (HCA) and geographic information system (GIS) in an attempt to discriminate the sources of the variation of water quality variations. The results of PCA technique indicate that variables responsible for water quality composition are mainly related to soluble salts variables; natural processes and the nature of the rock which modifies significantly the water chemistry. Inferred from the positive correlation between K+ and NO3-, NO3- is believed to be human induced rather than naturally originated. In this study, the multivariate statistical analysis and GIS allows the hydrogeologist to have supplementary tools in the characterization and evaluating of aquifers.Keywords: cluster, analysis, GIS, groundwater, laghouat, quality
Procedia PDF Downloads 32627864 Flood Mapping Using Height above the Nearest Drainage Model: A Case Study in Fredericton, NB, Canada
Authors: Morteza Esfandiari, Shabnam Jabari, Heather MacGrath, David Coleman
Abstract:
Flood is a severe issue in different places in the world as well as the city of Fredericton, New Brunswick, Canada. The downtown area of Fredericton is close to the Saint John River, which is susceptible to flood around May every year. Recently, the frequency of flooding seems to be increased, especially after the fact that the downtown area and surrounding urban/agricultural lands got flooded in two consecutive years in 2018 and 2019. In order to have an explicit vision of flood span and damage to affected areas, it is necessary to use either flood inundation modelling or satellite data. Due to contingent availability and weather dependency of optical satellites, and limited existing data for the high cost of hydrodynamic models, it is not always feasible to rely on these sources of data to generate quality flood maps after or during the catastrophe. Height Above the Nearest Drainage (HAND), a state-of-the-art topo-hydrological index, normalizes the height of a basin based on the relative elevation along with the stream network and specifies the gravitational or the relative drainage potential of an area. HAND is a relative height difference between the stream network and each cell on a Digital Terrain Model (DTM). The stream layer is provided through a multi-step, time-consuming process which does not always result in an optimal representation of the river centerline depending on the topographic complexity of that region. HAND is used in numerous case studies with quite acceptable and sometimes unexpected results because of natural and human-made features on the surface of the earth. Some of these features might cause a disturbance in the generated model, and consequently, the model might not be able to predict the flow simulation accurately. We propose to include a previously existing stream layer generated by the province of New Brunswick and benefit from culvert maps to improve the water flow simulation and accordingly the accuracy of HAND model. By considering these parameters in our processing, we were able to increase the accuracy of the model from nearly 74% to almost 92%. The improved model can be used for generating highly accurate flood maps, which is necessary for future urban planning and flood damage estimation without any need for satellite imagery or hydrodynamic computations.Keywords: HAND, DTM, rapid floodplain, simplified conceptual models
Procedia PDF Downloads 15327863 Adopted Method of Information System Strategy for Knowledge Management System: A Literature Review
Authors: Elin Cahyaningsih, Dana Indra Sensuse, Wahyu Catur Wibowo, Sofiyanti Indriasari
Abstract:
Bureaucracy reform program drives Indonesian government to change their management and supporting unit in order to enhance their organization performance. Information technology as one of supporting unit became one of strategic plan that organization tried to improve, because IT can automate and speed up process, reduce business process life cycle become more effective and efficient. Knowledge management system is a technology application for supporting knowledge management implementation in government which is requirement based on problem and potential functionality of each knowledge management process. Define knowledge management that suitable for each organization it is difficult, that why we should make the knowledge management system strategy as an alignment of knowledge management process in the organization. Knowledge management system is one of information system development in people perspective, because this system has high dependency in human interaction and participation. Strategic plan for developing knowledge management system can be determine using some of information system strategic methods. This research conducted to define type of strategic method of information system, stage of activity each method, the strategic method strength and weakness. The author use literature review methods for identify and classify strategic methods of information system for differentiate method type, categorize common activities, strength and weakness. Result of this research are determine and compare six strategic information system methods, there are Balanced Scorecard, Five Force Porter, SWOT analysis, Value Chain Analysis, Risk Analysis and Gap Analysis. Balanced Scorecard and Risk Analysis believe as common strategic method that usually used and have the highest excellence strength.Keywords: knowledge management system, balanced scorecard, five force, risk analysis, gap analysis, value chain analysis, SWOT analysis
Procedia PDF Downloads 48127862 Phrasemes With The Component 'Water' In Polish And Russian - Comparative Aspects
Authors: Aleksandra Majewska
Abstract:
The subject of this article is phrasemes with the component 'water' in Polish and Russian. The purpose of the study is to analyse the collocations from the point of view of lexis and semantics. The material for analysis was extracted from phraseological dictionaries of Polish and Russian. From the point of view of lexis, an analysis was made of the inflectional component 'water' in phrasal expressions in both languages. Then, the phrasemes were divided into their corresponding semantic groups. That division became the subject of another comparative analysis in a further step. Finally, the functioning of some phrasemes compounds in the contexts of modern Polish and Russian was shown.Keywords: lingustic, language, phraseme, polish and Russian
Procedia PDF Downloads 4427861 Antioxidant Face Mask from Purple Sweet Potato (Ipomea Batatas) with Oleum Cytrus
Authors: Lilis Kistriyani, Dine Olisvia, Lutfa Rahmawati
Abstract:
Facial mask is an important part of every beauty treatment because it will give a smooth and gentle effect on the face. This research is done to make edible film that will be applied for face mask. The main ingredient in making this edible film is purple sweet potato powder with the addition of glycerol as plasticizer. One of the ingredients in purple sweet potato is a flavonoid compound. The purpose of this study was to determine the effect of increasing the amount of glycerol to flavonoids release and the effect on the physical properties and biological properties of edible film produced. The stages of this research are the making of edible film, then perform some analysis, among others, spectrophotometer UV-vis analysis to find out how many flavonoids can be released into facial skin, tensile strength and elongation of break analysis, biodegradability analysis, and microbiological analysis. The variation of edible film is the volume of glycerol that is 1 ml, 2 ml, 3 ml. The results of spectrophotometer UV-vis analysis showed that the most flavonoid release concentration is 20.33 ppm in the 2 ml glycerol variation. The best tensile strength value is 8,502 N, and the greatest elongation of break value is 14% in 1 ml glycerol variation. In the biodegradability test, the more volume of glycerol added the faster the edible film is degraded. The results of microbiological analysis showed that purple sweet potato extract has the ability to inhibit the growth of Propionibacterium acnes seen in the presence of inhibiting zone which is 18.9 mm.Keywords: face mask, edible film, plasticizer, flavonoid
Procedia PDF Downloads 17827860 User Requirements Analysis for the Development of Assistive Navigation Mobile Apps for Blind and Visually Impaired People
Authors: Paraskevi Theodorou, Apostolos Meliones
Abstract:
In the context of the development process of two assistive navigation mobile apps for blind and visually impaired people (BVI) an extensive qualitative analysis of the requirements of potential users has been conducted. The analysis was based on interviews with BVIs and aimed to elicit not only their needs with respect to autonomous navigation but also their preferences on specific features of the apps under development. The elicited requirements were structured into four main categories, namely, requirements concerning the capabilities, functionality and usability of the apps, as well as compatibility requirements with respect to other apps and services. The main categories were then further divided into nine sub-categories. This classification, along with its content, aims to become a useful tool for the researcher or the developer who is involved in the development of digital services for BVI.Keywords: accessibility, assistive mobile apps, blind and visually impaired people, user requirements analysis
Procedia PDF Downloads 12527859 Review of Cable Fault Locating Methods and Usage of VLF for Real Cases of High Resistance Fault Locating
Authors: Saadat Ali, Rashid Abdulla Ahmed Alshehhi
Abstract:
Cable faults are always probable and common during or after commissioning, causing significant delays and disrupting power distribution or transmission network, which is intolerable for the utilities&service providers being their reliability and business continuity measures. Therefore, the adoption of rapid localization & rectification methodology is the main concern for them. This paper explores the present techniques available for high voltage cable localization & rectification and which is preferable with regards to easier, faster, and also less harmful to cables. It also provides insight experience of high resistance fault locating by utilization of the Very Low Frequency (VLF) method.Keywords: faults, VLF, real cases, cables
Procedia PDF Downloads 11327858 The Analysis of Brain Response to Auditory Stimuli through EEG Signals’ Non-Linear Analysis
Authors: H. Namazi, H. T. N. Kuan
Abstract:
Brain activity can be measured by acquiring and analyzing EEG signals from an individual. In fact, the human brain response to external and internal stimuli is mapped in his EEG signals. During years some methods such as Fourier transform, wavelet transform, empirical mode decomposition, etc. have been used to analyze the EEG signals in order to find the effect of stimuli, especially external stimuli. But each of these methods has some weak points in analysis of EEG signals. For instance, Fourier transform and wavelet transform methods are linear signal analysis methods which are not good to be used for analysis of EEG signals as nonlinear signals. In this research we analyze the brain response to auditory stimuli by extracting information in the form of various measures from EEG signals using a software developed by our research group. The used measures are Jeffrey’s measure, Fractal dimension and Hurst exponent. The results of these analyses are useful not only for fundamental understanding of brain response to auditory stimuli but provide us with very good recommendations for clinical purposes.Keywords: auditory stimuli, brain response, EEG signal, fractal dimension, hurst exponent, Jeffrey’s measure
Procedia PDF Downloads 53427857 Design, Development by Functional Analysis in UML and Static Test of a Multimedia Voice and Video Communication Platform on IP for a Use Adapted to the Context of Local Businesses in Lubumbashi
Authors: Blaise Fyama, Elie Museng, Grace Mukoma
Abstract:
In this article we present a java implementation of video telephony using the SIP protocol (Session Initiation Protocol). After a functional analysis of the SIP protocol, we relied on the work of Italian researchers of University of Parma-Italy to acquire adequate libraries for the development of our own communication tool. In order to optimize the code and improve the prototype, we used, in an incremental approach, test techniques based on a static analysis based on the evaluation of the complexity of the software with the application of metrics and the number cyclomatic of Mccabe. The objective is to promote the emergence of local start-ups producing IP video in a well understood local context. We have arrived at the creation of a video telephony tool whose code is optimized.Keywords: static analysis, coding complexity metric mccabe, Sip, uml
Procedia PDF Downloads 12127856 Using a Character’s Inner Monologue for Song Analysis
Authors: Robert Roznowski
Abstract:
The thought process of the character is never more evident than when singing alone onstage. The composer scores the emotional state and the lyricist voices the inner conflict as the character shares with an audience her or his deepest feelings. It is at these moments that a character may be thought of as voicing her or his inner monologue. Using examples from several musical theatre songs, this presentation will look at a codified approach to analyze a song from a more psychological perspective. Using the clues from the score, traditional character analysis and a psychological-based scoring method an actor may explore more fully inhabit and express the sung and unsung thoughts of the character. The approach yields a richer and more complex approach to acting the song.Keywords: acting, analysis, musical theatre, psychology
Procedia PDF Downloads 48027855 Differentiation between Different Rangeland Sites Using Principal Component Analysis in Semi-Arid Areas of Sudan
Authors: Nancy Ibrahim Abdalla, Abdelaziz Karamalla Gaiballa
Abstract:
Rangelands in semi-arid areas provide a good source for feeding huge numbers of animals and serving environmental, economic and social importance; therefore, these areas are considered economically very important for the pastoral sector in Sudan. This paper investigates the means of differentiating between different rangelands sites according to soil types using principal component analysis to assist in monitoring and assessment purposes. Three rangeland sites were identified in the study area as flat sandy sites, sand dune site, and hard clay site. Principal component analysis (PCA) was used to reduce the number of factors needed to distinguish between rangeland sites and produce a new set of data including the most useful spectral information to run satellite image processing. It was performed using selected types of data (two vegetation indices, topographic data and vegetation surface reflectance within the three bands of MODIS data). Analysis with PCA indicated that there is a relatively high correspondence between vegetation and soil of the total variance in the data set. The results showed that the use of the principal component analysis (PCA) with the selected variables showed a high difference, reflected in the variance and eigenvalues and it can be used for differentiation between different range sites.Keywords: principal component analysis, PCA, rangeland sites, semi-arid areas, soil types
Procedia PDF Downloads 19027854 Two Degree of Freedom Spherical Mechanism Design for Exact Sun Tracking
Authors: Osman Acar
Abstract:
Sun tracking systems are the systems following the sun ray by a right angle or by predetermined certain angle. In this study, we used theoretical trajectory of sun for latitude of central Anatolia in Turkey. A two degree of freedom spherical mechanism was designed to have a large workspace able to follow the sun's theoretical motion by the right angle during the whole year. An inverse kinematic analysis was generated to find the positions of mechanism links for the predicted trajectory. Force and torque analysis were shown for the first day of the year.Keywords: sun tracking, theoretical sun trajectory, spherical mechanism, inverse kinematic analysis
Procedia PDF Downloads 42027853 Optimization Analysis of a Concentric Tube Heat Exchanger with Field Synergy Principle
Abstract:
The paper investigates the optimization analysis to the heat exchanger design, mainly with response surface method and genetic algorithm to explore the relationship between optimal fluid flow velocity and temperature of the heat exchanger using field synergy principle. First, finite volume method is proposed to calculate the flow temperature and flow rate distribution for numerical analysis. We identify the most suitable simulation equations by response surface methodology. Furthermore, a genetic algorithm approach is applied to optimize the relationship between fluid flow velocity and flow temperature of the heat exchanger. The results show that the field synergy angle plays vital role in the performance of a true heat exchanger.Keywords: optimization analysis, field synergy, heat exchanger, genetic algorithm
Procedia PDF Downloads 31027852 Techno-Economic Analysis of Solar Energy for Cathodic Protection of Oil and Gas Buried Pipelines in Southwestern of Iran
Authors: M. Goodarzi, M. Mohammadi, A. Gharib
Abstract:
Solar energy is a renewable energy which has attracted special attention in many countries. Solar cathodic protectionsystems harness the sun’senergy to protect underground pipelinesand tanks from galvanic corrosion. The object of this study is to design and the economic analysis a cathodic protection system by impressed current supplied with solar energy panels applied to underground pipelines. In the present study, the technical and economic analysis of using solar energy for cathodic protection system in southwestern of Iran (Khuzestan province) is investigated. For this purpose, the ecological conditions such as the weather data, air clearness and sunshine hours are analyzed. The economic analyses were done using computer code to investigate the feasibility analysis from the using of various energy sources in order to cathodic protection system. The overall research methodology is divided into four components: Data collection, design of elements, techno economical evaluation, and output analysis. According to the results, solar renewable energy systems can supply adequate power for cathodic protection system purposes.Keywords: renewable energy, solar energy, solar cathodic protection station, lifecycle cost method
Procedia PDF Downloads 54627851 Integrated Dynamic Analysis of Semi-Submersible Flap Type Concept
Authors: M. Rafiur Rahman, M. Mezbah Uddin, Mohammad Irfan Uddin, M. Moinul Islam
Abstract:
With a rapid development of offshore renewable energy industry, the research activities in regards of harnessing power from offshore wind and wave energy are increasing day by day. Integration of wind turbines and wave energy converters into one combined semi-submersible platform might be a cost-economy and beneficial option. In this paper, the coupled integrated dynamic analysis in the time domain (TD) of a simplified semi-submersible flap type concept (SFC) is accomplished via state-of-the-art numerical code referred as Simo-Riflex-Aerodyn (SRA). This concept is a combined platform consisting of a semi-submersible floater supporting a 5 MW horizontal axis wind turbine (WT) and three elliptical shaped flap type wave energy converters (WECs) on three pontoons. The main focus is to validate the numerical model of SFC with experimental results and perform the frequency domain (FD) and TD response analysis. The numerical analysis is performed using potential flow theory for hydrodynamics and blade element momentum (BEM) theory for aerodynamics. A variety of environmental conditions encompassing the functional & survival conditions for short-term sea (1-hour simulation) are tested to evaluate the sustainability of the SFC. The numerical analysis is performed in full scale. Finally, the time domain analysis of heave, pitch & surge motions is performed numerically using SRA and compared with the experimental results. Due to the simplification of the model, there are some discrepancies which are discussed in brief.Keywords: coupled integrated dynamic analysis, SFC, time domain analysis, wave energy converters
Procedia PDF Downloads 22427850 Dynamic Analysis of Double Deck Tunnel
Authors: C. W. Kwak, I. J. Park, D. I. Jang
Abstract:
The importance of cost-wise effective application and construction is getting increase due to the surge of traffic volume in the metropolitan cities. Accordingly, the necessity of the tunnel has large section becomes more critical. Double deck tunnel can be one of the most appropriate solutions to the necessity. The dynamic stability of double deck tunnel is essential against seismic load since it has large section and connection between perimeter lining and interim slab. In this study, 3-dimensional dynamic numerical analysis was conducted based on the Finite Difference Method to investigate the seismic behavior of double deck tunnel. Seismic joint for dynamic stability and the mitigation of seismic impact on the lining was considered in the modeling and analysis. Consequently, the mitigation of acceleration, lining displacement and stress were verified successfully.Keywords: double deck tunnel, interim slab, 3-dimensional dynamic numerical analysis, seismic joint
Procedia PDF Downloads 38227849 A Bibliometric Analysis of Research on E-learning in Physics Education: Trends, Patterns, and Future Directions
Authors: Siti Nurjanah, Supahar
Abstract:
E-learning has become an increasingly popular mode of instruction, particularly in the field of physics education, where it offers opportunities for interactive and engaging learning experiences. This research aims to analyze the trends of research that investigated e-learning in physics education. Data was extracted from Scopus's database using the keywords "physics" and "e-learning". Of the 380 articles obtained based on the search criteria, a trend analysis of the research was carried out with the help of RStudio using the biblioshiny package and VosViewer software. Analysis showed that publications on this topic have increased significantly from 2014 to 2021. The publication was dominated by researchers from the United States. The main journal that publishes articles on this topic is Proceedings Frontiers in Education Conference fie. The most widely cited articles generally focus on the effectiveness of Moodle for physics learning. Overall, this research provides an in-depth understanding of the trends and key findings of research related to e-learning in physics.Keywords: bibliometric analysis, physics education, biblioshiny, E-learning
Procedia PDF Downloads 4427848 Clinical Feature Analysis and Prediction on Recurrence in Cervical Cancer
Authors: Ravinder Bahl, Jamini Sharma
Abstract:
The paper demonstrates analysis of the cervical cancer based on a probabilistic model. It involves technique for classification and prediction by recognizing typical and diagnostically most important test features relating to cervical cancer. The main contributions of the research include predicting the probability of recurrences in no recurrence (first time detection) cases. The combination of the conventional statistical and machine learning tools is applied for the analysis. Experimental study with real data demonstrates the feasibility and potential of the proposed approach for the said cause.Keywords: cervical cancer, recurrence, no recurrence, probabilistic, classification, prediction, machine learning
Procedia PDF Downloads 36027847 Application of Two Stages Adaptive Neuro-Fuzzy Inference System to Improve Dissolved Gas Analysis Interpretation Techniques
Authors: Kharisma Utomo Mulyodinoto, Suwarno, A. Abu-Siada
Abstract:
Dissolved Gas Analysis is one of impressive technique to detect and predict internal fault of transformers by using gas generated by transformer oil sample. A number of methods are used to interpret the dissolved gas from transformer oil sample: Doernenberg Ratio Method, IEC (International Electrotechnical Commission) Ratio Method, and Duval Triangle Method. While the assessment of dissolved gas within transformer oil samples has been standardized over the past two decades, analysis of the results is not always straight forward as it depends on personnel expertise more than mathematical formulas. To get over this limitation, this paper is aimed at improving the interpretation of Doernenberg Ratio Method, IEC Ratio Method, and Duval Triangle Method using Two Stages Adaptive Neuro-Fuzzy Inference System (ANFIS). Dissolved gas analysis data from 520 faulty transformers was analyzed to establish the proposed ANFIS model. Results show that the developed ANFIS model is accurate and can standardize the dissolved gas interpretation process with accuracy higher than 90%.Keywords: ANFIS, dissolved gas analysis, Doernenberg ratio method, Duval triangular method, IEC ratio method, transformer
Procedia PDF Downloads 15027846 Wet Spun Graphene Fibers With Silver Nanoparticles For Flexible Electronic Applications
Authors: Syed W. Hasan, Zhiqun Tian
Abstract:
Wet spinning provides a facile and economic route to fabricate graphene nanofibers (GFs) on mass scale. Nevertheless, the pristine GFs exhibit significantly low electrical and mechanical properties owing to stacked graphene sheets and weak inter-atomic bonding. In this report, we present highly conductive Ag-decorated-GFs (Ag/GFs). The SEM micrographs show Ag nanoparticles (NPs) (dia ~10 nm) are homogeneously distributed throughout the cross-section of the fiber. The Ag NPs provide a conductive network for the electrons flow raising the conductivity to 1.8(10^4) S/m which is 4 times higher than the pristine GFs. Our results surpass the conductivities of graphene fibers doped with CNTs, Nanocarbon, fullerene, and Cu. The chemical and structural attributes of Ag/GFs are further elucidated through XPS, AFM and Raman spectroscopy.Keywords: Ag nanoparticles, Conductive fibers, Graphene, Wet spinning
Procedia PDF Downloads 14427845 Failure Analysis Using Rtds for a Power System Equipped with Thyristor-Controlled Series Capacitor in Korea
Authors: Chur Hee Lee, Jae in Lee, Minh Chau Diah, Jong Su Yoon, Seung Wan Kim
Abstract:
This paper deals with Real Time Digital Simulator (RTDS) analysis about effects of transmission lines failure in power system equipped with Thyristor Controlled Series Capacitance (TCSC) in Korea. The TCSC is firstly applied in Korea to compensate real power in case of 765 kV line faults. Therefore, It is important to analyze with TCSC replica using RTDS. In this test, all systems in Korea, other than those near TCSC, were abbreviated to Thevenin equivalent. The replica was tested in the case of a line failure near the TCSC, a generator failure, and a 765-kV line failure. The effects of conventional operated STATCOM, SVC and TCSC were also analyzed. The test results will be used for the actual TCSC operational impact analysis.Keywords: failure analysis, power system, RTDS, TCSC
Procedia PDF Downloads 12227844 Relating Symptoms with Protein Production Abnormality in Patients with Down Syndrome
Authors: Ruolan Zhou
Abstract:
Trisomy of human chromosome 21 is the primary cause of Down Syndrome (DS), and this genetic disease has significantly burdened families and countries, causing great controversy. To address this problem, the research takes an approach in exploring the relationship between genetic abnormality and this disease's symptoms, adopting several techniques, including data analysis and enrichment analysis. It also explores open-source websites, such as NCBI, DAVID, SOURCE, STRING, as well as UCSC, to complement its result. This research has analyzed the variety of genes on human chromosome 21 with simple coding, and by using analysis, it has specified the protein-coding genes, their function, and their location. By using enrichment analysis, this paper has found the abundance of keratin production-related coding-proteins on human chromosome 21. By adopting past researches, this research has attempted to disclose the relationship between trisomy of human chromosome 21 and keratin production abnormality, which might be the reason for common diseases in patients with Down Syndrome. At last, by addressing the advantage and insufficiency of this research, the discussion has provided specific directions for future research.Keywords: Down Syndrome, protein production, genome, enrichment analysis
Procedia PDF Downloads 12827843 Hybrid Knowledge and Data-Driven Neural Networks for Diffuse Optical Tomography Reconstruction in Medical Imaging
Authors: Paola Causin, Andrea Aspri, Alessandro Benfenati
Abstract:
Diffuse Optical Tomography (DOT) is an emergent medical imaging technique which employs NIR light to estimate the spatial distribution of optical coefficients in biological tissues for diagnostic purposes, in a noninvasive and non-ionizing manner. DOT reconstruction is a severely ill-conditioned problem due to prevalent scattering of light in the tissue. In this contribution, we present our research in adopting hybrid knowledgedriven/data-driven approaches which exploit the existence of well assessed physical models and build upon them neural networks integrating the availability of data. Namely, since in this context regularization procedures are mandatory to obtain a reasonable reconstruction [1], we explore the use of neural networks as tools to include prior information on the solution. 2. Materials and Methods The idea underlying our approach is to leverage neural networks to solve PDE-constrained inverse problems of the form 𝒒 ∗ = 𝒂𝒓𝒈 𝒎𝒊𝒏𝒒 𝐃(𝒚, 𝒚̃), (1) where D is a loss function which typically contains a discrepancy measure (or data fidelity) term plus other possible ad-hoc designed terms enforcing specific constraints. In the context of inverse problems like (1), one seeks the optimal set of physical parameters q, given the set of observations y. Moreover, 𝑦̃ is the computable approximation of y, which may be as well obtained from a neural network but also in a classic way via the resolution of a PDE with given input coefficients (forward problem, Fig.1 box ). Due to the severe ill conditioning of the reconstruction problem, we adopt a two-fold approach: i) we restrict the solutions (optical coefficients) to lie in a lower-dimensional subspace generated by auto-decoder type networks. This procedure forms priors of the solution (Fig.1 box ); ii) we use regularization procedures of type 𝒒̂ ∗ = 𝒂𝒓𝒈𝒎𝒊𝒏𝒒 𝐃(𝒚, 𝒚̃)+ 𝑹(𝒒), where 𝑹(𝒒) is a regularization functional depending on regularization parameters which can be fixed a-priori or learned via a neural network in a data-driven modality. To further improve the generalizability of the proposed framework, we also infuse physics knowledge via soft penalty constraints (Fig.1 box ) in the overall optimization procedure (Fig.1 box ). 3. Discussion and Conclusion DOT reconstruction is severely hindered by ill-conditioning. The combined use of data-driven and knowledgedriven elements is beneficial and allows to obtain improved results, especially with a restricted dataset and in presence of variable sources of noise.Keywords: inverse problem in tomography, deep learning, diffuse optical tomography, regularization
Procedia PDF Downloads 7527842 Laban Movement Analysis Using Kinect
Authors: Bernstein Ran, Shafir Tal, Tsachor Rachelle, Studd Karen, Schuster Assaf
Abstract:
Laban Movement Analysis (LMA), developed in the dance community over the past seventy years, is an effective method for observing, describing, notating, and interpreting human movement to enhance communication and expression in everyday and professional life. Many applications that use motion capture data might be significantly leveraged if the Laban qualities will be recognized automatically. This paper presents an automated recognition method of Laban qualities from motion capture skeletal recordings and it is demonstrated on the output of Microsoft’s Kinect V2 sensor.Keywords: Laban movement analysis, multitask learning, Kinect sensor, machine learning
Procedia PDF Downloads 34427841 Generalized Additive Model for Estimating Propensity Score
Authors: Tahmidul Islam
Abstract:
Propensity Score Matching (PSM) technique has been widely used for estimating causal effect of treatment in observational studies. One major step of implementing PSM is estimating the propensity score (PS). Logistic regression model with additive linear terms of covariates is most used technique in many studies. Logistics regression model is also used with cubic splines for retaining flexibility in the model. However, choosing the functional form of the logistic regression model has been a question since the effectiveness of PSM depends on how accurately the PS been estimated. In many situations, the linearity assumption of linear logistic regression may not hold and non-linear relation between the logit and the covariates may be appropriate. One can estimate PS using machine learning techniques such as random forest, neural network etc for more accuracy in non-linear situation. In this study, an attempt has been made to compare the efficacy of Generalized Additive Model (GAM) in various linear and non-linear settings and compare its performance with usual logistic regression. GAM is a non-parametric technique where functional form of the covariates can be unspecified and a flexible regression model can be fitted. In this study various simple and complex models have been considered for treatment under several situations (small/large sample, low/high number of treatment units) and examined which method leads to more covariate balance in the matched dataset. It is found that logistic regression model is impressively robust against inclusion quadratic and interaction terms and reduces mean difference in treatment and control set equally efficiently as GAM does. GAM provided no significantly better covariate balance than logistic regression in both simple and complex models. The analysis also suggests that larger proportion of controls than treatment units leads to better balance for both of the methods.Keywords: accuracy, covariate balances, generalized additive model, logistic regression, non-linearity, propensity score matching
Procedia PDF Downloads 369