Search results for: exponential smoothing methods
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15661

Search results for: exponential smoothing methods

15241 Vocal Training and Practice Methods: A Glimpse on the South Indian Carnatic Music

Authors: Raghavi Janaswamy, Saraswathi K. Vasudev

Abstract:

Music is one of the supreme arts of expressions, next to the speech itself. Its evolution over centuries has paved the way with a variety of training protocols and performing methods. Indian classical music is one of the most elaborate and refined systems with immense emphasis on the voice culture related to range, breath control, quality of the tone, flexibility and diction. Several exercises namely saraliswaram, jantaswaram, dhatuswaram, upper stayi swaram, alamkaras and varnams lay the required foundation to gain the voice culture and deeper understanding on the voice development and further on to the intricacies of the raga system. This article narrates a few of the Carnatic music training methods with an emphasis on the advanced practice methods for articulating the vocal skills, continuity in the voice, ability to produce gamakams, command in the multiple speeds of rendering with reasonable volume. The creativity on these exercises and their impact on the voice production are discussed. The articulation of the outlined conscious practice methods and vocal exercises bestow the optimum use of the natural human vocal system to not only enhance the signing quality but also to gain health benefits.

Keywords: Carnatic music, Saraliswaram, Varnam, vocal training

Procedia PDF Downloads 176
15240 Preparation of MgO Nanoparticles by Green Methods

Authors: Maryam Sabbaghan, Pegah Sofalgar

Abstract:

Over the past few decades, a significant amount of research activities in the chemical community has been directed towards green synthesis. This area of chemistry has received extensive attention because of environmentally benign processes as well as economically viable. In this article, the MgO nanoparticles were prepared by different methods in the present of ionic liquids. A wide range of Magnesium oxide particle sizes within the nanometer scale is obtained by these methods. The structure of these MgO particles was studied by using X-ray diffraction (XRD), Infrared spectroscopy (IR), and scanning electron microscopy (SEM). It was found that the formation of nanoparticle could involve the role of performed 'nucleus' and used template to control the growth rate of nucleuses. The crystallite size of the MgO products was in a range from 31 to 77 nm.

Keywords: MgO, ionic liquid, nanoparticles, green chemistry

Procedia PDF Downloads 290
15239 A Survey on Concurrency Control Methods in Distributed Database

Authors: Seyed Mohsen Jameii

Abstract:

In the last years, remarkable improvements have been made in the ability of distributed database systems performance. A distributed database is composed of some sites which are connected to each other through network connections. In this system, if good harmonization is not made between different transactions, it may result in database incoherence. Nowadays, because of the complexity of many sites and their connection methods, it is difficult to extend different models in distributed database serially. The principle goal of concurrency control in distributed database is to ensure not interfering in accessibility of common database by different sites. Different concurrency control algorithms have been suggested to use in distributed database systems. In this paper, some available methods have been introduced and compared for concurrency control in distributed database.

Keywords: distributed database, two phase locking protocol, transaction, concurrency

Procedia PDF Downloads 352
15238 Zinc Oxide Varistor Performance: A 3D Network Model

Authors: Benjamin Kaufmann, Michael Hofstätter, Nadine Raidl, Peter Supancic

Abstract:

ZnO varistors are the leading overvoltage protection elements in today’s electronic industry. Their highly non-linear current-voltage characteristics, very fast response times, good reliability and attractive cost of production are unique in this field. There are challenges and questions unsolved. Especially, the urge to create even smaller, versatile and reliable parts, that fit industry’s demands, brings manufacturers to the limits of their abilities. Although, the varistor effect of sintered ZnO is known since the 1960’s, and a lot of work was done on this field to explain the sudden exponential increase of conductivity, the strict dependency on sinter parameters, as well as the influence of the complex microstructure, is not sufficiently understood. For further enhancement and down-scaling of varistors, a better understanding of the microscopic processes is needed. This work attempts a microscopic approach to investigate ZnO varistor performance. In order to cope with the polycrystalline varistor ceramic and in order to account for all possible current paths through the material, a preferably realistic model of the microstructure was set up in the form of three-dimensional networks where every grain has a constant electric potential, and voltage drop occurs only at the grain boundaries. The electro-thermal workload, depending on different grain size distributions, was investigated as well as the influence of the metal-semiconductor contact between the electrodes and the ZnO grains. A number of experimental methods are used, firstly, to feed the simulations with realistic parameters and, secondly, to verify the obtained results. These methods are: a micro 4-point probes method system (M4PPS) to investigate the current-voltage characteristics between single ZnO grains and between ZnO grains and the metal electrode inside the varistor, micro lock-in infrared thermography (MLIRT) to detect current paths, electron back scattering diffraction and piezoresponse force microscopy to determine grain orientations, atom probe to determine atomic substituents, Kelvin probe force microscopy for investigating grain surface potentials. The simulations showed that, within a critical voltage range, the current flow is localized along paths which represent only a tiny part of the available volume. This effect could be observed via MLIRT. Furthermore, the simulations exhibit that the electric power density, which is inversely proportional to the number of active current paths, since this number determines the electrical active volume, is dependent on the grain size distribution. M4PPS measurements showed that the electrode-grain contacts behave like Schottky diodes and are crucial for asymmetric current path development. Furthermore, evaluation of actual data suggests that current flow is influenced by grain orientations. The present results deepen the knowledge of influencing microscopic factors on ZnO varistor performance and can give some recommendations on fabrication for obtaining more reliable ZnO varistors.

Keywords: metal-semiconductor contact, Schottky diode, varistor, zinc oxide

Procedia PDF Downloads 281
15237 Design of Active Power Filters for Harmonics on Power System and Reducing Harmonic Currents

Authors: Düzgün Akmaz, Hüseyin Erişti

Abstract:

In the last few years, harmonics have been occurred with the increasing use of nonlinear loads, and these harmonics have been an ever increasing problem for the line systems. This situation importantly affects the quality of power and gives large losses to the network. An efficient way to solve these problems is providing harmonic compensation through parallel active power filters. Many methods can be used in the control systems of the parallel active power filters which provide the compensation. These methods efficiently affect the performance of the active power filters. For this reason, the chosen control method is significant. In this study, Fourier analysis (FA) control method and synchronous reference frame (SRF) control method are discussed. These control methods are designed for both eliminate harmonics and perform reactive power compensation in MATLAB/Simulink pack program and are tested. The results have been compared for each two methods.

Keywords: parallel active power filters, harmonic compensation, power quality, harmonics

Procedia PDF Downloads 459
15236 A Homogenized Mechanical Model of Carbon Nanotubes/Polymer Composite with Interface Debonding

Authors: Wenya Shu, Ilinca Stanciulescu

Abstract:

Carbon nanotubes (CNTs) possess attractive properties, such as high stiffness and strength, and high thermal and electrical conductivities, making them promising filler in multifunctional nanocomposites. Although CNTs can be efficient reinforcements, the expected level of mechanical performance of CNT-polymers is not often reached in practice due to the poor mechanical behavior of the CNT-polymer interfaces. It is believed that the interactions of CNT and polymer mainly result from the Van der Waals force. The interface debonding is a fracture and delamination phenomenon. Thus, the cohesive zone modeling (CZM) is deemed to give good capture of the interface behavior. The detailed, cohesive zone modeling provides an option to consider the CNT-matrix interactions, but brings difficulties in mesh generation and also leads to high computational costs. Homogenized models that smear the fibers in the ground matrix and treat the material as homogeneous are studied in many researches to simplify simulations. But based on the perfect interface assumption, the traditional homogenized model obtained by mixing rules severely overestimates the stiffness of the composite, even comparing with the result of the CZM with artificially very strong interface. A mechanical model that can take into account the interface debonding and achieve comparable accuracy to the CZM is thus essential. The present study first investigates the CNT-matrix interactions by employing cohesive zone modeling. Three different coupled CZM laws, i.e., bilinear, exponential and polynomial, are considered. These studies indicate that the shapes of the CZM constitutive laws chosen do not influence significantly the simulations of interface debonding. Assuming a bilinear traction-separation relationship, the debonding process of single CNT in the matrix is divided into three phases and described by differential equations. The analytical solutions corresponding to these phases are derived. A homogenized model is then developed by introducing a parameter characterizing interface sliding into the mixing theory. The proposed mechanical model is implemented in FEAP8.5 as a user material. The accuracy and limitations of the model are discussed through several numerical examples. The CZM simulations in this study reveal important factors in the modeling of CNT-matrix interactions. The analytical solutions and proposed homogenized model provide alternative methods to efficiently investigate the mechanical behaviors of CNT/polymer composites.

Keywords: carbon nanotube, cohesive zone modeling, homogenized model, interface debonding

Procedia PDF Downloads 129
15235 Conceptual and Funnel Methods Contribution to Critical Literature Review: PhD construction Management

Authors: Samuel Quashie

Abstract:

This study is aimed at demonstrating the applicability and contribution of ‘Conceptual and Funnelling Methods’ during the literature review stages, for PhD in Construction Management, which focused on the ‘Development of an Integrated Management for Post-Disaster Reconstruction’, the viability of this approach using conceptual and funnel methods are demonstrated. The ‘conceptual review method’ builds upon the strengths of relevant material, detailing major points and areas covered and evaluates lesser relevant literature. Publications are reviewed in an integrated style, challenging the scientific theory and seeking to develop new insights. The funnel method grouped reviews by commonality, regardless of the topic or thesis statement. It shows that the literature review is acquired using different kinds of information to increase the variety and diversity of the investigation. Results demonstrated conceptual and funnel methods ability to reviewed and appraised the relevant literature. It puts them into an integrated style, allows an evaluation of credentials, originality, theory base, context and significance of the quality work to emerge. Objectives of the review are met and gaps in knowledge are identified and direct further studies to answer the research questions.

Keywords: Ph.D, construction management, critical literature review, conceptual and funnel methods

Procedia PDF Downloads 415
15234 Classifying Time Independent Plane Symmetric Spacetime through Noether`s Approach

Authors: Nazish Iftikhar, Adil Jhangeer, Tayyaba Naz

Abstract:

The universe is expanding at an accelerated rate. Symmetries are useful in understanding universe’s behavior. Emmy Noether reported the relation between symmetries and conservation laws. These symmetries are known as Noether symmetries which correspond to a conserved quantity. In differential equations, conservation laws play an important role. Noether symmetries are helpful in modified theories of gravity. Time independent plane symmetric spacetime was classified by Noether`s theorem. By using Noether`s theorem, set of linear partial differential equations was obtained having A(r), B(r) and F(r) as unknown radial functions. The Lagrangian corresponding to considered spacetime in the Noether equation was used to get Noether operators. Different possibilities of radial functions were considered. Firstly, all functions were same. All the functions were considered as non-zero constant, linear, reciprocal and exponential respectively. Secondly, two functions were proportional to each other keeping third function different. Second case has four subcases in which four different relationships between A(r), B(r) and F(r) were discussed. In all cases, we obtained nontrivial Noether operators including gauge term. Conserved quantities for each Noether operators were also presented.

Keywords: Noether gauge symmetries, radial function, Noether operator, conserved quantities

Procedia PDF Downloads 230
15233 Similarity Based Membership of Elements to Uncertain Concept in Information System

Authors: M. Kamel El-Sayed

Abstract:

The process of determining the degree of membership for an element to an uncertain concept has been found in many ways, using equivalence and symmetry relations in information systems. In the case of similarity, these methods did not take into account the degree of symmetry between elements. In this paper, we use a new definition for finding the membership based on the degree of symmetry. We provide an example to clarify the suggested methods and compare it with previous methods. This method opens the door to more accurate decisions in information systems.

Keywords: information system, uncertain concept, membership function, similarity relation, degree of similarity

Procedia PDF Downloads 223
15232 Review and Evaluation of Trending Canonical Correlation Analyses-Based Brain Computer Interface Methods

Authors: Bayar Shahab

Abstract:

The fast development of technology that has advanced neuroscience and human interaction with computers has enabled solutions to various problems, and issues of this new era have been found and are being found like no other time in history. Brain-computer interface so-called BCI has opened the door to several new research areas and have been able to provide solutions to critical and important issues such as supporting a paralyzed patient to interact with the outside world, controlling a robot arm, playing games in VR with the brain, driving a wheelchair or even a car and neurotechnology enabled the rehabilitation of the lost memory, etc. This review work presents state-of-the-art methods and improvements of canonical correlation analyses (CCA), which is an SSVEP-based BCI method. These are the methods used to extract EEG signal features or, to be said in a different way, the features of interest that we are looking for in the EEG analyses. Each of the methods from oldest to newest has been discussed while comparing their advantages and disadvantages. This would create a great context and help researchers to understand the most state-of-the-art methods available in this field with their pros and cons, along with their mathematical representations and usage. This work makes a vital contribution to the existing field of study. It differs from other similar recently published works by providing the following: (1) stating most of the prominent methods used in this field in a hierarchical way (2) explaining pros and cons of each method and their performance (3) presenting the gaps that exist at the end of each method that can open the understanding and doors to new research and/or improvements.

Keywords: BCI, CCA, SSVEP, EEG

Procedia PDF Downloads 145
15231 TessPy – Spatial Tessellation Made Easy

Authors: Jonas Hamann, Siavash Saki, Tobias Hagen

Abstract:

Discretization of urban areas is a crucial aspect in many spatial analyses. The process of discretization of space into subspaces without overlaps and gaps is called tessellation. It helps understanding spatial space and provides a framework for analyzing geospatial data. Tessellation methods can be divided into two groups: regular tessellations and irregular tessellations. While regular tessellation methods, like squares-grids or hexagons-grids, are suitable for addressing pure geometry problems, they cannot take the unique characteristics of different subareas into account. However, irregular tessellation methods allow the border between the subareas to be defined more realistically based on urban features like a road network or Points of Interest (POI). Even though Python is one of the most used programming languages when it comes to spatial analysis, there is currently no library that combines different tessellation methods to enable users and researchers to compare different techniques. To close this gap, we are proposing TessPy, an open-source Python package, which combines all above-mentioned tessellation methods and makes them easily accessible to everyone. The core functions of TessPy represent the five different tessellation methods: squares, hexagons, adaptive squares, Voronoi polygons, and city blocks. By using regular methods, users can set the resolution of the tessellation which defines the finesse of the discretization and the desired number of tiles. Irregular tessellation methods allow users to define which spatial data to consider (e.g., amenity, building, office) and how fine the tessellation should be. The spatial data used is open-source and provided by OpenStreetMap. This data can be easily extracted and used for further analyses. Besides the methodology of the different techniques, the state-of-the-art, including examples and future work, will be discussed. All dependencies can be installed using conda or pip; however, the former is more recommended.

Keywords: geospatial data science, geospatial data analysis, tessellations, urban studies

Procedia PDF Downloads 128
15230 A Comparative Study of Generalized Autoregressive Conditional Heteroskedasticity (GARCH) and Extreme Value Theory (EVT) Model in Modeling Value-at-Risk (VaR)

Authors: Longqing Li

Abstract:

The paper addresses the inefficiency of the classical model in measuring the Value-at-Risk (VaR) using a normal distribution or a Student’s t distribution. Specifically, the paper focuses on the one day ahead Value-at-Risk (VaR) of major stock market’s daily returns in US, UK, China and Hong Kong in the most recent ten years under 95% confidence level. To improve the predictable power and search for the best performing model, the paper proposes using two leading alternatives, Extreme Value Theory (EVT) and a family of GARCH models, and compares the relative performance. The main contribution could be summarized in two aspects. First, the paper extends the GARCH family model by incorporating EGARCH and TGARCH to shed light on the difference between each in estimating one day ahead Value-at-Risk (VaR). Second, to account for the non-normality in the distribution of financial markets, the paper applies Generalized Error Distribution (GED), instead of the normal distribution, to govern the innovation term. A dynamic back-testing procedure is employed to assess the performance of each model, a family of GARCH and the conditional EVT. The conclusion is that Exponential GARCH yields the best estimate in out-of-sample one day ahead Value-at-Risk (VaR) forecasting. Moreover, the discrepancy of performance between the GARCH and the conditional EVT is indistinguishable.

Keywords: Value-at-Risk, Extreme Value Theory, conditional EVT, backtesting

Procedia PDF Downloads 321
15229 Weighted Risk Scores Method Proposal for Occupational Safety Risk Assessment

Authors: Ulas Cinar, Omer Faruk Ugurlu, Selcuk Cebi

Abstract:

Occupational safety risk management is the most important element of a safe working environment. Effective risk management can only be possible with accurate analysis and evaluations. Scoring-based risk assessment methods offer considerable ease of application as they convert linguistic expressions into numerical results. It can also be easily adapted to any field. Contrary to all these advantages, important problems in scoring-based methods are frequently discussed. Effective measurability is one of the most critical problems. Existing methods allow experts to choose a score equivalent to each parameter. Therefore, experts prefer the score of the most likely outcome for risk. However, all other possible consequences are neglected. Assessments of the existing methods express the most probable level of risk, not the real risk of the enterprises. In this study, it is aimed to develop a method that will present a more comprehensive evaluation compared to the existing methods by evaluating the probability and severity scores, all sub-parameters, and potential results, and a new scoring-based method is proposed in the literature.

Keywords: occupational health and safety, risk assessment, scoring based risk assessment method, underground mining, weighted risk scores

Procedia PDF Downloads 135
15228 Assessment of Digital Literacy Skills of Librarians in Tertiary Institutions Inniger State

Authors: Mustapha Abdulkadir Gana, Jibrin Attahiru Alhassan, Adamu Musa Baba

Abstract:

The exponential growth of information sources, resources and the continued Communication Technology (ICT) sophistication of libraries all over the world call for capable and ICT compliant librarians in Nigeria, this article assesses the digital literacy skills of librarians in tertiary institutions in Niger state. The survey research method was applied in the study using a random sampling technique to draw the sample. Fifty-eight copies of the questionnaire were administered while forty-nine copies were completed, returned, and used in the study, which represents 84% of the response rate. Two research questions were answered, and data were analyzed using Statistical Package for the Social Sciences (SPSS). The finding uncovered that the librarians lack the requisite digital literacy skills to access the wealth of digital information resources available. The study recommends some steps to turn around the situations amongst; librarians must be empowered with all necessary digital literacy skills, embark on rigorous training and retraining programs, workshops, conferences, and seminars, there should also be a coherent training policy for the librarians on a sustainable basis to increase their requisite digital literacy skills.

Keywords: digital, information, literacy, skills

Procedia PDF Downloads 148
15227 Application of extraction chromatography to the separation of Sc, Zr and Sn isotopes from target materials

Authors: Steffen Happel

Abstract:

Non-standard isotopes such as Sc-44/47, Zr-89, and Sn-117mare finding interest is increasing in radiopharmaceutical applications. Methods for the separation of these elements from typical target materials were developed. The methods used in this paper are based on the use of extraction chromatographic resins such as UTEVA, TBP, and DGA resin. Information on the selectivity of the resins (Dw values of selected elements in HCl and HNO3 of varying concentration) will be presented as well as results of the method development such as elution studies, chemical recoveries, and decontamination factors. Developed methods are based on the use of vacuum supported separation allowing for fast and selective separation.

Keywords: elution, extraction chromatography, radiopharmacy, decontamination factors

Procedia PDF Downloads 468
15226 Analysis of the Aquifer Vulnerability of a Miopliocene Arid Area Using Drastic and SI Models

Authors: H. Majour, L. Djabri

Abstract:

Many methods in the groundwater vulnerability have been developed in the world (methods like PRAST, DRIST, APRON/ARAA, PRASTCHIM, GOD). In this study, our choice dealt with two recent complementary methods using category mapping of index with weighting criteria (Point County Systems Model MSCP) namely the standard DRASTIC method and SI (Susceptibility Index). At present, these two methods are the most used for the mapping of the intrinsic vulnerability of groundwater. Two classes of groundwater vulnerability in the Biskra sandy aquifer were identified by the DRASTIC method (average and high) and the SI method (very high and high). Integrated analysis has revealed that the high class is predominant for the DRASTIC method whereas for that of SI the preponderance is for the very high class. Furthermore, we notice that the method SI estimates better the vulnerability for the pollution in nitrates, with a rate of 85 % between the concentrations in nitrates of groundwater and the various established classes of vulnerability, against 75 % for the DRASTIC method. By including the land use parameter, the SI method produced more realistic results.

Keywords: DRASTIC, SI, GIS, Biskra sandy aquifer, Algeria

Procedia PDF Downloads 486
15225 Off-Line Detection of "Pannon Wheat" Milling Fractions by Near-Infrared Spectroscopic Methods

Authors: E. Izsó, M. Bartalné-Berceli, Sz. Gergely, A. Salgó

Abstract:

The aims of this investigation is to elaborate near-infrared methods for testing and recognition of chemical components and quality in “Pannon wheat” allied (i.e. true to variety or variety identified) milling fractions as well as to develop spectroscopic methods following the milling processes and evaluate the stability of the milling technology by different types of milling products and according to sampling times, respectively. This wheat categories produced under industrial conditions where samples were collected versus sampling time and maximum or minimum yields. The changes of the main chemical components (such as starch, protein, lipid) and physical properties of fractions (particle size) were analysed by dispersive spectrophotometers using visible (VIS) and near-infrared (NIR) regions of the electromagnetic radiation. Close correlation were obtained between the data of spectroscopic measurement techniques processed by various chemometric methods (e.g. principal component analysis (PCA), cluster analysis (CA) and operation condition of milling technology. Its obvious that NIR methods are able to detect the deviation of the yield parameters and differences of the sampling times by a wide variety of fractions, respectively. NIR technology can be used in the sensitive monitoring of milling technology.

Keywords: near infrared spectroscopy, wheat categories, milling process, monitoring

Procedia PDF Downloads 406
15224 Damage Assessment and Repair for Older Brick Buildings

Authors: Tim D. Sass

Abstract:

The experience of engineers and architects practicing today is typically limited to current building code requirements and modern construction methods and materials. However, many cities have a mix of new and old buildings with many buildings constructed over one hundred years ago when building codes and construction methods were much different. When a brick building sustains damage, a structural engineer is often hired to determine the cause of damage as well as determine the necessary repairs. Forensic studies of dozens of brick buildings shows an appreciation of historical building methods and materials is needed to correctly identify the cause of damage and design an appropriate repair. Damage on an older, brick building can be mistakenly attributed to storms or seismic events when the real source of the damage is deficient original construction. Assessing and remediating damaged brickwork on older brick buildings requires an understanding of the original construction, an understanding of older repair methods, and, an understanding of current building code requirements.

Keywords: brick, damage, deterioration, facade

Procedia PDF Downloads 227
15223 Chemical Hazards Impact on Efficiency of Energy Storage Battery and its Possible Mitigation's

Authors: Abirham Simeneh Ayalew, Seada Hussen Adem, Frie Ayalew Yimam

Abstract:

Battery energy storage has a great role on storing energy harnessed from different alternative resources and greatly benefit the power sector by supply energy back to the system during outage and regular operation in power sectors. Most of the study shows that there is an exponential increase in the quantity of lithium - ion battery energy storage system due to their power density, economical aspects and its performance. But this lithium ion battery failures resulted in fire and explosion due to its having flammable electrolytes (chemicals) which can create those hazards. Hazards happen in these energy storage system lead to minimize battery life spans or efficiency. Identifying the real cause of these hazards and its mitigation techniques can be the solution to improve the efficiency of battery technologies and the electrode materials should have high electrical conductivity, large surface area, stable structure and low resistance. This paper asses the real causes of chemical hazards, its impact on efficiency, proposed solution for mitigating those hazards associated with efficiency improvement and summery of researchers new finding related to the field.

Keywords: battery energy storage, battery energy storage efficiency, chemical hazards, lithium ion battery

Procedia PDF Downloads 78
15222 Comparative Evaluation of EBT3 Film Dosimetry Using Flat Bad Scanner, Densitometer and Spectrophotometer Methods and Its Applications in Radiotherapy

Authors: K. Khaerunnisa, D. Ryangga, S. A. Pawiro

Abstract:

Over the past few decades, film dosimetry has become a tool which is used in various radiotherapy modalities, either for clinical quality assurance (QA) or dose verification. The response of the film to irradiation is usually expressed in optical density (OD) or net optical density (netOD). While the film's response to radiation is not linear, then the use of film as a dosimeter must go through a calibration process. This study aimed to compare the function of the calibration curve of various measurement methods with various densitometer, using a flat bad scanner, point densitometer and spectrophotometer. For every response function, a radichromic film calibration curve is generated from each method by performing accuracy, precision and sensitivity analysis. netOD is obtained by measuring changes in the optical density (OD) of the film before irradiation and after irradiation when using a film scanner if it uses ImageJ to extract the pixel value of the film on the red channel of three channels (RGB), calculate the change in OD before and after irradiation when using a point densitometer, and calculate changes in absorbance before and after irradiation when using a spectrophotometer. the results showed that the three calibration methods gave readings with a netOD precision of doses below 3% for the uncertainty value of 1σ (one sigma). while the sensitivity of all three methods has the same trend in responding to film readings against radiation, it has a different magnitude of sensitivity. while the accuracy of the three methods provides readings below 3% for doses above 100 cGy and 200 cGy, but for doses below 100 cGy found above 3% when using point densitometers and spectrophotometers. when all three methods are used for clinical implementation, the results of the study show accuracy and precision below 2% for the use of scanners and spectrophotometers and above 3% for precision and accuracy when using point densitometers.

Keywords: Callibration Methods, Film Dosimetry EBT3, Flat Bad Scanner, Densitomete, Spectrophotometer

Procedia PDF Downloads 134
15221 Impact of Increasing Distributed Solar PV Systems on Distribution Networks in South Africa

Authors: Aradhna Pandarum

Abstract:

South Africa is experiencing an exponential growth of distributed solar PV installations. This is due to various factors with the predominant one being increasing electricity tariffs along with decreasing installation costs, resulting in attractive business cases to some end-users. Despite there being a variety of economic and environmental advantages associated with the installation of PV, their potential impact on distribution grids has yet to be thoroughly investigated. This is especially true since the locations of these units cannot be controlled by Network Service Providers (NSPs) and their output power is stochastic and non-dispatchable. This report details two case studies that were completed to determine the possible voltage and technical losses impact of increasing PV penetration in the Northern Cape of South Africa. Some major impacts considered for the simulations were ramping of PV generation due to intermittency caused by moving clouds, the size and overall hosting capacity and the location of the systems. The main finding is that the technical impact is different on a constrained feeder vs a non-constrained feeder. The acceptable PV penetration level is much lower for a constrained feeder than a non-constrained feeder, depending on where the systems are located.

Keywords: medium voltage networks, power system losses, power system voltage, solar photovoltaic

Procedia PDF Downloads 153
15220 Breast Cancer Early Recognition, New Methods of Screening, and Analysis

Authors: Sahar Heidary

Abstract:

Breast cancer is a main public common obstacle global. Additionally, it is the second top reason for tumor death across women. Considering breast cancer cure choices can aid private doctors in precaution for their patients through future cancer treatment. This article reviews usual management centered on stage, histology, and biomarkers. The growth of breast cancer is a multi-stage procedure including numerous cell kinds and its inhibition residues stimulating in the universe. Timely identification of breast cancer is one of the finest methods to stop this illness. Entirely chief therapeutic administrations mention screening mammography for women aged 40 years and older. Breast cancer metastasis interpretations for the mainstream of deaths from breast cancer. The discovery of breast cancer metastasis at the initial step is essential for managing and estimate of breast cancer development. Developing methods consuming the exploration of flowing cancer cells illustrate talented outcomes in forecasting and classifying the initial steps of breast cancer metastasis in patients. In public, mammography residues are the key screening implement though the efficiency of medical breast checks and self-checkup is less. Innovative screening methods are doubtful to exchange mammography in the close upcoming for screening the overall people.

Keywords: breast cancer, screening, metastasis, methods

Procedia PDF Downloads 167
15219 A Review of Travel Data Collection Methods

Authors: Muhammad Awais Shafique, Eiji Hato

Abstract:

Household trip data is of crucial importance for managing present transportation infrastructure as well as to plan and design future facilities. It also provides basis for new policies implemented under Transportation Demand Management. The methods used for household trip data collection have changed with passage of time, starting with the conventional face-to-face interviews or paper-and-pencil interviews and reaching to the recent approach of employing smartphones. This study summarizes the step-wise evolution in the travel data collection methods. It provides a comprehensive review of the topic, for readers interested to know the changing trends in the data collection field.

Keywords: computer, smartphone, telephone, travel survey

Procedia PDF Downloads 313
15218 Physical and Chemical Alternative Methods of Fresh Produce Disinfection

Authors: Tuji Jemal Ahmed

Abstract:

Fresh produce is an essential component of a healthy diet. However, it can also be a potential source of pathogenic microorganisms that can cause foodborne illnesses. Traditional disinfection methods, such as washing with water and chlorine, have limitations and may not effectively remove or inactivate all microorganisms. This has led to the development of alternative/new methods of fresh produce disinfection, including physical and chemical methods. In this paper, we explore the physical and chemical new methods of fresh produce disinfection, their advantages and disadvantages, and their suitability for different types of produce. Physical methods of disinfection, such as ultraviolet (UV) radiation and high-pressure processing (HPP), are crucial in ensuring the microbiological safety of fresh produce. UV radiation uses short-wavelength UV-C light to damage the DNA and RNA of microorganisms, and HPP applies high levels of pressure to fresh produce to reduce the microbial load. These physical methods are highly effective in killing a wide range of microorganisms, including bacteria, viruses, and fungi. However, they may not penetrate deep enough into the product to kill all microorganisms and can alter the sensory characteristics of the product. Chemical methods of disinfection, such as acidic electrolyzed water (AEW), ozone, and peroxyacetic acid (PAA), are also important in ensuring the microbiological safety of fresh produce. AEW uses a low concentration of hypochlorous acid and a high concentration of hydrogen ions to inactivate microorganisms, ozone uses ozone gas to damage the cell membranes and DNA of microorganisms, and PAA uses a combination of hydrogen peroxide and acetic acid to inactivate microorganisms. These chemical methods are highly effective in killing a wide range of microorganisms, but they may cause discoloration or changes in the texture and flavor of some products and may require specialized equipment and trained personnel to produce and apply. In conclusion, the selection of the most suitable method of fresh produce disinfection should take into consideration the type of product, the level of microbial contamination, the effectiveness of the method in reducing the microbial load, and any potential negative impacts on the sensory characteristics, nutritional composition, and safety of the produce.

Keywords: fresh produce, pathogenic microorganisms, foodborne illnesses, disinfection methods

Procedia PDF Downloads 73
15217 A Comparative Study of Medical Image Segmentation Methods for Tumor Detection

Authors: Mayssa Bensalah, Atef Boujelben, Mouna Baklouti, Mohamed Abid

Abstract:

Image segmentation has a fundamental role in analysis and interpretation for many applications. The automated segmentation of organs and tissues throughout the body using computed imaging has been rapidly increasing. Indeed, it represents one of the most important parts of clinical diagnostic tools. In this paper, we discuss a thorough literature review of recent methods of tumour segmentation from medical images which are briefly explained with the recent contribution of various researchers. This study was followed by comparing these methods in order to define new directions to develop and improve the performance of the segmentation of the tumour area from medical images.

Keywords: features extraction, image segmentation, medical images, tumor detection

Procedia PDF Downloads 167
15216 Comparison between Deterministic and Probabilistic Stability Analysis, Featuring Consequent Risk Assessment

Authors: Isabela Moreira Queiroz

Abstract:

Slope stability analyses are largely carried out by deterministic methods and evaluated through a single security factor. Although it is known that the geotechnical parameters can present great dispersal, such analyses are considered fixed and known. The probabilistic methods, in turn, incorporate the variability of input key parameters (random variables), resulting in a range of values of safety factors, thus enabling the determination of the probability of failure, which is an essential parameter in the calculation of the risk (probability multiplied by the consequence of the event). Among the probabilistic methods, there are three frequently used methods in geotechnical society: FOSM (First-Order, Second-Moment), Rosenblueth (Point Estimates) and Monte Carlo. This paper presents a comparison between the results from deterministic and probabilistic analyses (FOSM method, Monte Carlo and Rosenblueth) applied to a hypothetical slope. The end was held to evaluate the behavior of the slope and consequent risk analysis, which is used to calculate the risk and analyze their mitigation and control solutions. It can be observed that the results obtained by the three probabilistic methods were quite close. It should be noticed that the calculation of the risk makes it possible to list the priority to the implementation of mitigation measures. Therefore, it is recommended to do a good assessment of the geological-geotechnical model incorporating the uncertainty in viability, design, construction, operation and closure by means of risk management. 

Keywords: probabilistic methods, risk assessment, risk management, slope stability

Procedia PDF Downloads 391
15215 Structural Health Monitoring and Damage Structural Identification Using Dynamic Response

Authors: Reza Behboodian

Abstract:

Monitoring the structural health and diagnosing their damage in the early stages has always been one of the topics of concern. Nowadays, research on structural damage detection methods based on vibration analysis is very extensive. Moreover, these methods can be used as methods of permanent and timely inspection of structures and prevent further damage to structures. Non-destructive methods are the low-cost and economical methods for determining the damage of structures. In this research, a non-destructive method for detecting and identifying the failure location in structures based on dynamic responses resulting from time history analysis is proposed. When the structure is damaged due to the reduction of stiffness, and due to the applied loads, the displacements in different parts of the structure were increased. In the proposed method, the damage position is determined based on the calculation of the strain energy difference in each member of the damaged structure and the healthy structure at any time. Defective members of the structure are indicated by the amount of strain energy relative to the healthy state. The results indicated that the proper accuracy and performance of the proposed method for identifying failure in structures.

Keywords: failure, time history analysis, dynamic response, strain energy

Procedia PDF Downloads 133
15214 An Overview of Privacy and Security Issues in Social Networks

Authors: Mohamad Ibrahim Al Ladan

Abstract:

Social networks, such as Facebook, Myspace, LinkedIn, Google+, and Twitter have experienced exponential growth and a remarkable adoption rate in recent years. They provide attractive means of online social interactions and communications with family, friends, and colleagues from around the corner or across the globe, and they have become an important part of daily digital interactions for more than one and a half billion users around the world. The various personal information sharing practices that social network providers encourage have led to their success as innovative social interaction platforms. However, these practices have resulted in ample concerns with respect to privacy and security from different stakeholders. Addressing these privacy and security concerns in social networks is a must for these networks to be sustainable. Existing security and privacy tools may not be enough to address existing concerns. Some guidelines should be followed to protect users from the existing risks. In this paper, we have investigated and discussed the various privacy and security issues and concerns pertaining to social networks. Moreover, we have classified these privacy and security issues and presented a thorough discussion of the implications of these issues and concerns on the future of the social networks. In addition, we have presented a set of guidelines as precaution measures that users can consider to address these issues and concerns.

Keywords: social networks privacy issues, social networks security issues, social networks privacy precautions measures, social networks security precautions measures

Procedia PDF Downloads 307
15213 Optimization of Machine Learning Regression Results: An Application on Health Expenditures

Authors: Songul Cinaroglu

Abstract:

Machine learning regression methods are recommended as an alternative to classical regression methods in the existence of variables which are difficult to model. Data for health expenditure is typically non-normal and have a heavily skewed distribution. This study aims to compare machine learning regression methods by hyperparameter tuning to predict health expenditure per capita. A multiple regression model was conducted and performance results of Lasso Regression, Random Forest Regression and Support Vector Machine Regression recorded when different hyperparameters are assigned. Lambda (λ) value for Lasso Regression, number of trees for Random Forest Regression, epsilon (ε) value for Support Vector Regression was determined as hyperparameters. Study results performed by using 'k' fold cross validation changed from 5 to 50, indicate the difference between machine learning regression results in terms of R², RMSE and MAE values that are statistically significant (p < 0.001). Study results reveal that Random Forest Regression (R² ˃ 0.7500, RMSE ≤ 0.6000 ve MAE ≤ 0.4000) outperforms other machine learning regression methods. It is highly advisable to use machine learning regression methods for modelling health expenditures.

Keywords: machine learning, lasso regression, random forest regression, support vector regression, hyperparameter tuning, health expenditure

Procedia PDF Downloads 226
15212 Method to Create Signed Word - Application in Teaching and Learning Vietnamese Sign Language

Authors: Nguyen Thi Kim Thoa

Abstract:

Vietnam currently has about two million five hundred deaf/hard of hearing people. Although the issue of Vietnamese Sign Language (VSL) education has received attention from the State, there are still many issues that need to be resolved, such as policies, teacher training in both knowledge and teaching methods, education programs, and textbook compilation. Furthermore, the issue of research on VSL has not yet attracted the attention of linguists. Using the quantitative description method, the article will analyze, synthesize, and compare to find methods to create signed words in VSL, such as based on external shape characteristics, operational characteristics, operating methods, and basic meanings, from which we can see the special nature of signed words, the division of word types and the morphological meaning of creating new words through sign methods. From the results of this research, the aspect of ‘visual culture’ will be clarified in Vietnamese Deaf Culture. Through that, we also develop a number of vocabulary teaching methods (such as teaching vocabulary through a group of methods of forming signed words, teaching vocabulary using mind maps, and teaching vocabulary through culture...), with the aim of further improving the effectiveness of teaching and learning VSL in Vietnam. The research results also provide deaf people in Vietnam with a scientific and effective method of learning vocabulary, helping them quickly integrate into the community. The article will be a useful reference for linguists who want to research VSL.

Keywords: Vietnamese sign language (VSL), signed word, teaching, method

Procedia PDF Downloads 36