Search results for: statistical visualizing
237 Simulation and Statistical Analysis of Motion Behavior of a Single Rockfall
Authors: Iau-Teh Wang, Chin-Yu Lee
Abstract:
The impact force of a rockfall is mainly determined by its moving behavior and velocity, which are contingent on the rock shape, slope gradient, height, and surface roughness of the moving path. It is essential to precisely calculate the moving path of the rockfall in order to effectively minimize and prevent damages caused by the rockfall. By applying the Colorado Rockfall Simulation Program (CRSP) program as the analysis tool, this research studies the influence of three shapes of rock (spherical, cylindrical and discoidal) and surface roughness on the moving path of a single rockfall. As revealed in the analysis, in addition to the slope gradient, the geometry of the falling rock and joint roughness coefficient ( JRC ) of the slope are the main factors affecting the moving behavior of a rockfall. On a single flat slope, both the rock-s bounce height and moving velocity increase as the surface gradient increases, with a critical gradient value of 1:m = 1 . Bouncing behavior and faster moving velocity occur more easily when the rock geometry is more oval. A flat piece tends to cause sliding behavior and is easily influenced by the change of surface undulation. When JRC <1.4 the moving velocity decreases and the bounce height increases as JRC increases. If the gradient is fixed, when JRC is greater, the bounce height will be higher, while the moving velocity will experience a downward trend. Therefore, the best protecting point and facilities can be chosen if the moving paths of rockfalls are precisely estimated.Keywords: rock shape, surface roughness, moving path.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951236 The Effect of Cooperation Teaching Method on Learning of Students in Primary Schools
Authors: Fereshteh Afkari, Davood Bagheri
Abstract:
The effect of teaching method on learning assistance Dunn Review .The study, to compare the effects of collaboration on teaching mathematics learning courses, including writing, science, experimental girl students by other methods of teaching basic first paid and the amount of learning students methods have been trained to cooperate with other students with other traditional methods have been trained to compare. The survey on 100 students in Tehran that using random sampling ¬ cluster of girl students between the first primary selections was performed. Considering the topic of semi-experimental research methods used to practice the necessary information by questionnaire, examination questions by the researcher, in collaboration with teachers and view authority in this field and related courses that teach these must have been collected. Research samples to test and control groups were divided. Experimental group and control group collaboration using traditional methods of mathematics courses, including writing and experimental sciences were trained. Research results using statistical methods T is obtained in two independent groups show that, through training assistance will lead to positive results and student learning in comparison with traditional methods, will increase also led to collaboration methods increase skills to solve math lesson practice, better understanding and increased skill level of students in practical lessons such as science and has been writing.Keywords: method of teaching, learning, collaboration
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1637235 Governance Commitment and Time Differences in Aspects of Sustainability Reporting in Nigerian Banks
Authors: Nwobu Obiamaka, Owolabi Akintola
Abstract:
This study examined the extent of statistical significant difference between the economic, environmental, governance and social aspects of sustainability reporting as a result of board committee on sustainability and time (year) of reporting for business organizations in the Nigerian banking sector. The years of reporting under consideration were 2010, 2011, 2012 and 2013. Content analysis methodology was employed through a reporting index used to score the amount of economic, environmental, governance and social indicators of sustainability reporting. The results of this study indicated that business organizations with board committee on sustainability had more indicators of sustainability reporting than those without board committees on sustainability issues. Also, sustainability reporting in 2013 was higher than that of prior years (2012, 2011 and 2010) for the economic, environmental and social indicators. The governance indicators of 2012 was highest compared to the other years (2013, 2011 and 2010) under consideration in this study. The implication of this finding is that business organizations that have board committees on sustainability are monitored by such boards to report more to their stakeholders. On the other hand, business organizations are appreciating the need to engage in sustainability reporting with each passing year. This could be due to the Central Bank of Nigeria (CBN) Sustainability Reporting framework that business organizations in the banking sector have to adhere to. When sustainability issues are monitored from the board of directors, business organizations are likely to increase and improve on their sustainability reporting.Keywords: Governance, organizations, reporting, sustainability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2100234 Least Square-SVM Detector for Wireless BPSK in Multi-Environmental Noise
Authors: J. P. Dubois, Omar M. Abdul-Latif
Abstract:
Support Vector Machine (SVM) is a statistical learning tool developed to a more complex concept of structural risk minimization (SRM). In this paper, SVM is applied to signal detection in communication systems in the presence of channel noise in various environments in the form of Rayleigh fading, additive white Gaussian background noise (AWGN), and interference noise generalized as additive color Gaussian noise (ACGN). The structure and performance of SVM in terms of the bit error rate (BER) metric is derived and simulated for these advanced stochastic noise models and the computational complexity of the implementation, in terms of average computational time per bit, is also presented. The performance of SVM is then compared to conventional binary signaling optimal model-based detector driven by binary phase shift keying (BPSK) modulation. We show that the SVM performance is superior to that of conventional matched filter-, innovation filter-, and Wiener filter-driven detectors, even in the presence of random Doppler carrier deviation, especially for low SNR (signal-to-noise ratio) ranges. For large SNR, the performance of the SVM was similar to that of the classical detectors. However, the convergence between SVM and maximum likelihood detection occurred at a higher SNR as the noise environment became more hostile.Keywords: Colour noise, Doppler shift, innovation filter, least square-support vector machine, matched filter, Rayleigh fading, Wiener filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1813233 Cost of Governance in Nigeria: In Whose Interest?
Authors: Francis O. Iyoha, Daniel E. Gberevbie, Charles T. Iruonagbe, Matthew E. Egharevba
Abstract:
Cost of governance in Nigeria has become a challenge to development and concern to practitioners and scholars alike in the field of business and social science research. In the 2010 national budget of NGN4.6 trillion or USD28.75billion for instance, only a pantry sum of NGN1.8trillion or USD11.15billion was earmarked for capital expenditure. Similarly, in 2013, out of a total national budget of NGN4.92trillion or USD30.75billion, only the sum of NGN1.50trllion or USD9.38billion was voted for capital expenditure. Therefore, based on the data sourced from the Nigerian Office of Statistics, Central bank of Nigeria Statistical Bulletin as well as from the United Nations Development Programme, this study examined the causes of high cost of governance in Nigeria. It found out that the high cost of governance in the country is in the interest of the ruling class, arising from their unethical behaviour – corrupt practices and the poor management of public resources. As a result, the study recommends the need to intensify the war against corruption and mismanagement of public resources by government officials as possible solution to overcome the high cost of governance in Nigeria. This could be achieved by strengthening the constitutional powers of the various anti-corruption agencies in the area of arrest, investigation and prosecution of offenders without the interference of the executive arm of government either at the local, state or federal level.
Keywords: Capital expenditure, Cost of governance, recurrent expenditure, unethical behaviour.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3530232 AniMoveMineR: Animal Behavior Exploratory Analysis Using Association Rules Mining
Authors: Suelane Garcia Fontes, Silvio Luiz Stanzani, Pedro L. Pizzigatti Corrła Ronaldo G. Morato
Abstract:
Environmental changes and major natural disasters are most prevalent in the world due to the damage that humanity has caused to nature and these damages directly affect the lives of animals. Thus, the study of animal behavior and their interactions with the environment can provide knowledge that guides researchers and public agencies in preservation and conservation actions. Exploratory analysis of animal movement can determine the patterns of animal behavior and with technological advances the ability of animals to be tracked and, consequently, behavioral studies have been expanded. There is a lot of research on animal movement and behavior, but we note that a proposal that combines resources and allows for exploratory analysis of animal movement and provide statistical measures on individual animal behavior and its interaction with the environment is missing. The contribution of this paper is to present the framework AniMoveMineR, a unified solution that aggregates trajectory analysis and data mining techniques to explore animal movement data and provide a first step in responding questions about the animal individual behavior and their interactions with other animals over time and space. We evaluated the framework through the use of monitored jaguar data in the city of Miranda Pantanal, Brazil, in order to verify if the use of AniMoveMineR allows to identify the interaction level between these jaguars. The results were positive and provided indications about the individual behavior of jaguars and about which jaguars have the highest or lowest correlation.Keywords: Data mining, data science, trajectory, animal behavior.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 917231 Multi Task Scheme to Monitor Multivariate Environments Using Artificial Neural Network
Authors: K. Atashgar
Abstract:
When an assignable cause(s) manifests itself to a multivariate process and the process shifts to an out-of-control condition, a root-cause analysis should be initiated by quality engineers to identify and eliminate the assignable cause(s) affected the process. A root-cause analysis in a multivariate process is more complex compared to a univariate process. In the case of a process involved several correlated variables an effective root-cause analysis can be only experienced when it is possible to identify the required knowledge including the out-of-control condition, the change point, and the variable(s) responsible to the out-of-control condition, all simultaneously. Although literature addresses different schemes to monitor multivariate processes, one can find few scientific reports focused on all the required knowledge. To the best of the author’s knowledge this is the first time that a multi task model based on artificial neural network (ANN) is reported to monitor all the required knowledge at the same time for a multivariate process with more than two correlated quality characteristics. The performance of the proposed scheme is evaluated numerically when different step shifts affect the mean vector. Average run length is used to investigate the performance of the proposed multi task model. The simulated results indicate the multi task scheme performs all the required knowledge effectively.
Keywords: Artificial neural network, Multivariate process, Statistical process control, Change point.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1681230 The Virtual Container Yard: Identifying the Persuasive Factors in Container Interchange
Authors: L. Edirisinghe, Zhihong Jin, A. W. Wijeratne, R. Mudunkotuwa
Abstract:
The virtual container yard is an effective solution to the container inventory imbalance problem which is a global issue. It causes substantial cost to carriers, which inadvertently adds to the prices of consumer goods. The virtual container yard is rooted in the fundamentals of container interchange between carriers. If carriers opt to interchange their excess containers with those who are deficit, a substantial part of the empty reposition cost could be eliminated. Unlike in other types of ships, cargo cannot be directly loaded to a container ship. Slots and containers are supplementary components; thus, without containers, a carrier cannot ship cargo if the containers are not available and vice versa. Few decades ago, carriers recognized slot (the unit of space in a container ship) interchange as a viable solution for the imbalance of shipping space. Carriers interchange slots among them and it also increases the advantage of scale of economies in container shipping. Some of these service agreements between mega carriers have provisions to interchange containers too. However, the interchange mechanism is still not popular among carriers for containers. This is the paradox that prevails in the liner shipping industry. At present, carriers reposition their excess empty containers to areas where they are in demand. This research applied factor analysis statistical method. The paper reveals that five major components may influence the virtual container yard namely organisation, practice and culture, legal and environment, international nature, and marketing. There are 12 variables that may impact the virtual container yard, and these are explained in the paper.
Keywords: Virtual container yard, imbalance, management, inventory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 839229 Quality Parameters of Offset Printing Wastewater
Authors: Kiurski S. Jelena, Kecić S. Vesna, Aksentijević M. Snežana
Abstract:
Samples of tap and wastewater were collected in three offset printing facilities in Novi Sad, Serbia. Ten physicochemical parameters were analyzed within all collected samples: pH, conductivity, m - alkalinity, p - alkalinity, acidity, carbonate concentration, hydrogen carbonate concentration, active oxygen content, chloride concentration and total alkali content. All measurements were conducted using the standard analytical and instrumental methods. Comparing the obtained results for tap water and wastewater, a clear quality difference was noticeable, since all physicochemical parameters were significantly higher within wastewater samples. The study also involves the application of simple linear regression analysis on the obtained dataset. By using software package ORIGIN 5 the pH value was mutually correlated with other physicochemical parameters. Based on the obtained values of Pearson coefficient of determination a strong positive correlation between chloride concentration and pH (r = -0.943), as well as between acidity and pH (r = -0.855) was determined. In addition, statistically significant difference was obtained only between acidity and chloride concentration with pH values, since the values of parameter F (247.634 and 182.536) were higher than Fcritical (5.59). In this way, results of statistical analysis highlighted the most influential parameter of water contamination in offset printing, in the form of acidity and chloride concentration. The results showed that variable dependence could be represented by the general regression model: y = a0 + a1x+ k, which further resulted with matching graphic regressions.
Keywords: Pollution, printing industry, simple linear regression analysis, wastewater.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1674228 Factors Influencing the Success of Mobile Phone Entrepreneurs at Central Plaza
Authors: NattapongTecharattanased
Abstract:
The purpose of this research was to study the factors that influenced the success of mobile phone entrepreneurs at Central Plaza. The sample group included 187 entrepreneurs at Central Plaza. A questionnaire was utilized as a tool to collect data. Statistics used in this research included frequency, percentage, mean, and standard deviation. Independent- sample t- test, one way ANOVA, and multiple regression analysis. Data were analyzed by using Statistical Package for the Social Sciences.The findings disclosed that the majority of respondents were male between 25-40 years old, and held an undergraduate degree. The average income of respondents was between 15,001-25,000 baht. The majority of respondents had less than 5 years of working experience. In terms of personality, the findings revealed that expression and agreement were ranked at the highest level. Whereas, emotion stability, consciousness, open to new experience were ranked at high. From the hypotheses testing, the findings revealed that different genders had different success in their mobile phone business with different income from the last 6 months. However, difference in age, income, level of education, and experience affected the success in terms of income, number of customers, and overall success of business. Moreover, the factors of personalities included expression, agreement, emotion stability, consciousness, open to new experience, and competitive strategy. From the findings, these factors were able to predict mobile phone business success at 66.9 percent.
Keywords: Entrepreneur, Influencing Factors, Success, Mobile Phone Business.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1662227 Combining the Deep Neural Network with the K-Means for Traffic Accident Prediction
Authors: Celso L. Fernando, Toshio Yoshii, Takahiro Tsubota
Abstract:
Understanding the causes of a road accident and predicting their occurrence is key to prevent deaths and serious injuries from road accident events. Traditional statistical methods such as the Poisson and the Logistics regressions have been used to find the association of the traffic environmental factors with the accident occurred; recently, an artificial neural network, ANN, a computational technique that learns from historical data to make a more accurate prediction, has emerged. Although the ability to make accurate predictions, the ANN has difficulty dealing with highly unbalanced attribute patterns distribution in the training dataset; in such circumstances, the ANN treats the minority group as noise. However, in the real world data, the minority group is often the group of interest; e.g., in the road traffic accident data, the events of the accident are the group of interest. This study proposes a combination of the k-means with the ANN to improve the predictive ability of the neural network model by alleviating the effect of the unbalanced distribution of the attribute patterns in the training dataset. The results show that the proposed method improves the ability of the neural network to make a prediction on a highly unbalanced distributed attribute patterns dataset; however, on an even distributed attribute patterns dataset, the proposed method performs almost like a standard neural network.
Keywords: Accident risks estimation, artificial neural network, deep learning, K-mean, road safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 973226 Durability Study Partially Saturated Fly Ash Blended Cement Concrete
Authors: N. Shafiq, M. F. Nuruddin, S. C. Chin
Abstract:
This paper presents the experimental results of the investigation of various properties related to the durability and longterm performance of mortars made of Fly Ash blended cement, FA and Ordinary Portland cement, OPC. The properties that were investigated in an experimental program include; equilibration of specimen in different relative humidity, determination of total porosity, compressive strength, chloride permeability index, and electrical resistivity. Fly Ash blended cement mortar specimens exhibited 10% to 15% lower porosity when measured at equilibrium conditions in different relative humidities as compared to the specimens made of OPC mortar, which resulted in 6% to 8% higher compressive strength of FA blended cement mortar specimens. The effects of ambient relative humidity during sample equilibration on porosity and strength development were also studied. For specimens equilibrated in higher relative humidity conditions, such as 75%, the total porosity of different mortar specimens was between 35% to 50% less than the porosity of samples equilibrated in 12% relative humidity, consequently leading to higher compressive strengths of these specimens.A valid statistical correlation between values of compressive strength, porosity and the degree of saturation was obtained. Measured values of chloride permeability index of fly ash blended cement mortar were obtained as one fourth to one sixth of those measured for OPC mortar specimens, which indicates high resistance against chloride ion penetration in FA blended cement specimens, hence resulting in a highly durable mortar.Keywords: chloride permeability index, equilibrium condition, electrical resistivity, fly ash
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1942225 Identification of the Electronic City Application Obstacles in Iran
Authors: E. Asgharizadeh, M. Ajalli Geshlajoughi, S. R. Safavi Mirmahalleh
Abstract:
Amazing development of the information technology, communications and internet expansion as well as the requirements of the city managers to new ideas to run the city and higher participation of the citizens encourage us to complete the electronic city as soon as possible. The foundations of this electronic city are in information technology. People-s participation in metropolitan management is a crucial topic. Information technology does not impede this matter. It can ameliorate populace-s participation and better interactions between the citizens and the city managers. Citizens can proffer their ideas, beliefs and votes through digital mass media based upon the internet and computerization plexuses on the topical matters to receive appropriate replies and services. They can participate in urban projects by becoming cognizant of the city views. The most significant challenges are as follows: information and communicative management, altering citizens- views, as well as legal and office documents Electronic city obstacles have been identified in this research. The required data were forgathered through questionnaires to identify the barriers from a statistical community comprising specialists and practitioners of the ministry of information technology and communication, the municipality information technology organization. The conclusions demonstrate that the prioritized electronic city application barriers in Iran are as follows: The support quandaries (non-financial ones), behavioral, cultural and educational plights, the security, legal and license predicaments, the hardware, orismological and infrastructural curbs, the software and fiscal problems.Keywords: Electronic city, urban management, populace's participation, electronic government, electronic services, electronic organization, electronic infrastructure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1762224 Using Fuzzy Logic Decision Support System to Predict the Lifted Weight for Students at Weightlifting Class
Authors: Ahmed Abdulghani Taha, Mohammad Abdulghani Taha
Abstract:
This study aims at being acquainted with the using the body fat percentage (%BF) with body Mass Index (BMI) as input parameters in fuzzy logic decision support system to predict properly the lifted weight for students at weightlifting class lift according to his abilities instead of traditional manner. The sample included 53 male students (age = 21.38 ± 0.71 yrs, height (Hgt) = 173.17 ± 5.28 cm, body weight (BW) = 70.34 ± 7.87.6 kg, Body mass index (BMI) 23.42 ± 2.06 kg.m-2, fat mass (FM) = 9.96 ± 3.15 kg and fat percentage (% BF) = 13.98 ± 3.51 %.) experienced the weightlifting class as a credit and has variance at BW, Hgt and BMI and FM. BMI and % BF were taken as input parameters in FUZZY logic whereas the output parameter was the lifted weight (LW). There were statistical differences between LW values before and after using fuzzy logic (Diff 3.55± 2.21, P > 0.001). The percentages of the LW categories proposed by fuzzy logic were 3.77% of students to lift 1.0 fold of their bodies; 50.94% of students to lift 0.95 fold of their bodies; 33.96% of students to lift 0.9 fold of their bodies; 3.77% of students to lift 0.85 fold of their bodies and 7.55% of students to lift 0.8 fold of their bodies. The study concluded that the characteristic changes in body composition experienced by students when undergoing weightlifting could be utilized side by side with the Fuzzy logic decision support system to determine the proper workloads consistent with the abilities of students.Keywords: Fuzzy logic, body mass index, body fat percentage, weightlifting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1533223 Perception of Secondary Schools’ Students on Computer Education in Federal Capital Territory (FCT-Abuja), Nigeria
Authors: Salako Emmanuel Adekunle
Abstract:
Computer education is referred to as the knowledge and ability to use computers and related technology efficiently, with a range of skills covering levels from basic use to advance. Computer continues to make an ever-increasing impact on all aspect of human endeavours such as education. With numerous benefits of computer education, what are the insights of students on computer education? This study investigated the perception of senior secondary school students on computer education in Federal Capital Territory (FCT), Abuja, Nigeria. A sample of 7500 senior secondary schools students was involved in the study, one hundred (100) private and fifty (50) public schools within FCT. They were selected by using simple random sampling technique. A questionnaire [PSSSCEQ] was developed and validated through expert judgement and reliability coefficient of 0.84 was obtained. It was used to gather relevant data on computer education. Findings confirmed that the students in the FCT had positive perception on computer education. Some factors were identified that affect students’ perception on computer education. The null hypotheses were tested using t-test and ANOVA statistical analyses at 0.05 level of significance. Based on these findings, some recommendations were made which include competent teachers should be employed into all secondary schools. This will help students to acquire relevant knowledge in computer education, technological supports should be provided to all secondary schools; this will help the users (students) to solve specific problems in computer education and financial supports should be provided to procure computer facilities that will enhance the teaching and the learning of computer education.Keywords: Computer education, perception, secondary school, students.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4063222 A Metric-Set and Model Suggestion for Better Software Project Cost Estimation
Authors: Murat Ayyıldız, Oya Kalıpsız, Sırma Yavuz
Abstract:
Software project effort estimation is frequently seen as complex and expensive for individual software engineers. Software production is in a crisis. It suffers from excessive costs. Software production is often out of control. It has been suggested that software production is out of control because we do not measure. You cannot control what you cannot measure. During last decade, a number of researches on cost estimation have been conducted. The metric-set selection has a vital role in software cost estimation studies; its importance has been ignored especially in neural network based studies. In this study we have explored the reasons of those disappointing results and implemented different neural network models using augmented new metrics. The results obtained are compared with previous studies using traditional metrics. To be able to make comparisons, two types of data have been used. The first part of the data is taken from the Constructive Cost Model (COCOMO'81) which is commonly used in previous studies and the second part is collected according to new metrics in a leading international company in Turkey. The accuracy of the selected metrics and the data samples are verified using statistical techniques. The model presented here is based on Multi-Layer Perceptron (MLP). Another difficulty associated with the cost estimation studies is the fact that the data collection requires time and care. To make a more thorough use of the samples collected, k-fold, cross validation method is also implemented. It is concluded that, as long as an accurate and quantifiable set of metrics are defined and measured correctly, neural networks can be applied in software cost estimation studies with successKeywords: Software Metrics, Software Cost Estimation, Neural Network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1957221 Spatial-Temporal Clustering Characteristics of Dengue in the Northern Region of Sri Lanka, 2010-2013
Authors: Sumiko Anno, Keiji Imaoka, Takeo Tadono, Tamotsu Igarashi, Subramaniam Sivaganesh, Selvam Kannathasan, Vaithehi Kumaran, Sinnathamby Noble Surendran
Abstract:
Dengue outbreaks are affected by biological, ecological, socio-economic and demographic factors that vary over time and space. These factors have been examined separately and still require systematic clarification. The present study aimed to investigate the spatial-temporal clustering relationships between these factors and dengue outbreaks in the northern region of Sri Lanka. Remote sensing (RS) data gathered from a plurality of satellites were used to develop an index comprising rainfall, humidity and temperature data. RS data gathered by ALOS/AVNIR-2 were used to detect urbanization, and a digital land cover map was used to extract land cover information. Other data on relevant factors and dengue outbreaks were collected through institutions and extant databases. The analyzed RS data and databases were integrated into geographic information systems, enabling temporal analysis, spatial statistical analysis and space-time clustering analysis. Our present results showed that increases in the number of the combination of ecological factor and socio-economic and demographic factors with above the average or the presence contribute to significantly high rates of space-time dengue clusters.
Keywords: ALOS/AVNIR-2, Dengue, Space-time clustering analysis, Sri Lanka.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2284220 Further Development in Predicting Post-Earthquake Fire Ignition Hazard
Authors: Pegah Farshadmanesh, Jamshid Mohammadi, Mehdi Modares
Abstract:
In nearly all earthquakes of the past century that resulted in moderate to significant damage, the occurrence of postearthquake fire ignition (PEFI) has imposed a serious hazard and caused severe damage, especially in urban areas. In order to reduce the loss of life and property caused by post-earthquake fires, there is a crucial need for predictive models to estimate the PEFI risk. The parameters affecting PEFI risk can be categorized as: 1) factors influencing fire ignition in normal (non-earthquake) condition, including floor area, building category, ignitability, type of appliance, and prevention devices, and 2) earthquake related factors contributing to the PEFI risk, including building vulnerability and earthquake characteristics such as intensity, peak ground acceleration, and peak ground velocity. State-of-the-art statistical PEFI risk models are solely based on limited available earthquake data, and therefore they cannot predict the PEFI risk for areas with insufficient earthquake records since such records are needed in estimating the PEFI model parameters. In this paper, the correlation between normal condition ignition risk, peak ground acceleration, and PEFI risk is examined in an effort to offer a means for predicting post-earthquake ignition events. An illustrative example is presented to demonstrate how such correlation can be employed in a seismic area to predict PEFI hazard.Keywords: Fire risk, post-earthquake fire ignition (PEFI), risk management, seismicity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1361219 Using Field Indices of Rill and Gully in order to Erosion Estimating and Sediment Analysis (Case Study: Menderjan Watershed in Isfahan Province, Iran)
Authors: Masoud Nasri, Sadat Feiznia, Mohammad Jafari, Hasan Ahmadi
Abstract:
Today, incorrect use of lands and land use changes, excessive grazing, no suitable using of agricultural farms, plowing on steep slopes, road construct, building construct, mine excavation etc have been caused increasing of soil erosion and sediment yield. For erosion and sediment estimation one can use statistical and empirical methods. This needs to identify land unit map and the map of effective factors. However, these empirical methods are usually time consuming and do not give accurate estimation of erosion. In this study, we applied GIS techniques to estimate erosion and sediment of Menderjan watershed at upstream Zayandehrud river in center of Iran. Erosion faces at each land unit were defined on the basis of land use, geology and land unit map using GIS. The UTM coordinates of each erosion type that showed more erosion amounts such as rills and gullies were inserted in GIS using GPS data. The frequency of erosion indicators at each land unit, land use and their sediment yield of these indices were calculated. Also using tendency analysis of sediment yield changes in watershed outlet (Menderjan hydrometric gauge station), was calculated related parameters and estimation errors. The results of this study according to implemented watershed management projects can be used for more rapid and more accurate estimation of erosion than traditional methods. These results can also be used for regional erosion assessment and can be used for remote sensing image processing.Keywords: Erosion and sedimentation, Gully, Rill, GIS, GPS, Menderjan Watershed
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1908218 Statistical Modeling of Local Area Fading Channels Based on Triply Stochastic Filtered Marked Poisson Point Processes
Authors: Jihad S. Daba, J. P. Dubois
Abstract:
Fading noise degrades the performance of cellular communication, most notably in femto- and pico-cells in 3G and 4G systems. When the wireless channel consists of a small number of scattering paths, the statistics of fading noise is not analytically tractable and poses a serious challenge to developing closed canonical forms that can be analysed and used in the design of efficient and optimal receivers. In this context, noise is multiplicative and is referred to as stochastically local fading. In many analytical investigation of multiplicative noise, the exponential or Gamma statistics are invoked. More recent advances by the author of this paper utilized a Poisson modulated-weighted generalized Laguerre polynomials with controlling parameters and uncorrelated noise assumptions. In this paper, we investigate the statistics of multidiversity stochastically local area fading channel when the channel consists of randomly distributed Rayleigh and Rician scattering centers with a coherent Nakagami-distributed line of sight component and an underlying doubly stochastic Poisson process driven by a lognormal intensity. These combined statistics form a unifying triply stochastic filtered marked Poisson point process model.
Keywords: Cellular communication, femto- and pico-cells, stochastically local area fading channel, triply stochastic filtered marked Poisson point process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1345217 Instant Location Detection of Objects Moving at High-Speedin C-OTDR Monitoring Systems
Authors: Andrey V. Timofeev
Abstract:
The practical efficient approach is suggested to estimate the high-speed objects instant bounds in C-OTDR monitoring systems. In case of super-dynamic objects (trains, cars) is difficult to obtain the adequate estimate of the instantaneous object localization because of estimation lag. In other words, reliable estimation coordinates of monitored object requires taking some time for data observation collection by means of C-OTDR system, and only if the required sample volume will be collected the final decision could be issued. But it is contrary to requirements of many real applications. For example, in rail traffic management systems we need to get data of the dynamic objects localization in real time. The way to solve this problem is to use the set of statistical independent parameters of C-OTDR signals for obtaining the most reliable solution in real time. The parameters of this type we can call as «signaling parameters» (SP). There are several the SP’s which carry information about dynamic objects instant localization for each of COTDR channels. The problem is that some of these parameters are very sensitive to dynamics of seismoacoustic emission sources, but are non-stable. On the other hand, in case the SP is very stable it becomes insensitive as rule. This report contains describing of the method for SP’s co-processing which is designed to get the most effective dynamic objects localization estimates in the C-OTDR monitoring system framework.
Keywords: C-OTDR-system, co-processing of signaling parameters, high-speed objects localization, multichannel monitoring systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1907216 Investigating the Precipitation and Temperature Change Procedure in Zayanderood Watershed
Authors: Amir Gandomkar
Abstract:
Global warming and continental changes have been one of the people's issues in the recent years and its consequences have appeared in the most parts of the earth planet or will appear in the future. Temperature and Precipitation are two main parameters in climatology. Any changes in these two parameters in this region cause widespread changes in the ecosystem and its natural and humanistic structure. One of the important consequences of this procedure is change in surface and underground water resources. Zayanderood watershed basin which is the main central river in Iran has faced water shortage in the recent years and also it has resulted in drought in Gavkhuni swamp and the river itself. Managers and experts in provinces which are the Zayanderood water consumers believe that global warming; raining decrease and continental changes are the main reason of water decrease. By statistical investigation of annual Precipitation and 46 years temperature of internal and external areas of Zayanderood watershed basin's stations and by using Kendal-man method, Precipitation and temperature procedure changes have been analyzed in this basin. According to obtained results, there was not any noticeable decrease or increase procedure in Precipitation and annual temperature in the basin during this period. However, regarding to Precipitation, a noticeable decrease and increase have been observed in small part of western and some parts of eastern and southern basin, respectively. Furthermore, the investigation of annual temperature procedure has shown that a noticeable increase has been observed in some parts of western and eastern basin, and also a noticeable increasing procedure of temperature in the central parts of metropolitan Esfahan can be observed.Keywords: Zayanderood, Man_Kendal, Climate Change
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1480215 Classification of Business Models of Italian Bancassurance by Balance Sheet Indicators
Authors: Andrea Bellucci, Martina Tofi
Abstract:
The aim of paper is to analyze business models of bancassurance in Italy for life business. The life insurance business is very developed in the Italian market and banks branches have 80% of the market share. Given its maturity, the life insurance market needs to consolidate its organizational form to allow for the development of non-life business, which nowadays collects few premiums but represents a great opportunity to enlarge the market share of bancassurance using its strength in the distribution channel while the market share of independent agents is decreasing. Starting with the main business model of bancassurance for life business, this paper will analyze the performances of life companies in the Italian market by balance sheet indicators and by main discriminant variables of business models. The study will observe trends from 2013 to 2015 for the Italian market by exploiting a database managed by Associazione Nazionale delle Imprese di Assicurazione (ANIA). The applied approach is based on a bottom-up analysis starting with variables and indicators to define business models’ classification. The statistical classification algorithm proposed by Ward is employed to design business models’ profiles. Results from the analysis will be a representation of the main business models built by their profile related to indicators. In that way, an unsupervised analysis is developed that has the limit of its judgmental dimension based on research opinion, but it is possible to obtain a design of effective business models.
Keywords: Balance sheet indicators, Bancassurance, business models, ward algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1261214 Development and Acceptance of a Proposed Module for Enhancing the Reading and Writing Skills in Baybayin: The Traditional Writing System in the Philippines
Authors: Maria Venus G. Solares
Abstract:
The ancient Filipinos had their own spelling or alphabet that differed from the modern Roman alphabet brought by the Spaniards. It consists of seventeen letters, three vowels, and fourteen consonants and is called Baybayin. The Baybayin, a traditional writing system, is composed of characters that represent syllables. A proposal in the Philippine Congress to declare Baybayin as the national writing system inspired this study. The main objective of this study was to develop and assess the proposed module for enhancing the reading and writing skills in Baybayin of the students. The aim was to ensure the acceptability of the Baybayin using the proposed module and to meet the needs of students in developing their ability to read and write Baybayin through the module. A quasi-experimental research design was used in this study. The data were collected through the initial and final analysis of the students of Adamson University's ABM 1102 using convenient sampling techniques. Based on statistical analysis of data using weighted mean, standard deviation, and paired t-tests, the proposed module helped improve the students' literacy skills, and the response exercises in the proposed module changed the acceptability of the Baybayin in their minds. The study showed that there was an important difference in the scores of students before and after the use of the module. The students' response to the assessment of their reading and writing skills on Baybayin was highly acceptable. This study will help to develop the reading and writing skills of the students in Baybayin and to teach the Baybayin in response to the revival of a part of Philippine culture that has been long forgotten.
Keywords: Baybayin, proposed module, ancient writing, acceptability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 48213 The Role of the Accused’s Attorney in the Criminal Justice System of Iran, Mashhad 2014
Authors: Mahdi Karimi
Abstract:
One of the most basic standards of fair trial is the right to defense, hire an attorney and its presence in the hearing stages. On the one hand, based on the reason and justice, as the legal issues, particularly criminal affairs, become complicated, the accused must benefit from an attorney in the court in order to defend itself which requires legal knowledge. On the other hand, as the judicial system has jurists such as investigation judges at its disposal, the accused must enjoy the same right to defend itself and reject allegations so that the balance is maintained between the litigating parties based on the principle of "equality of arms". The right to adequate time and facilities for defense is cited among the principles and rights relevant to the proceedings in international regulations such as the International Covenant on Civil and Political Rights. The innovations made in the Code of Criminal Procedure in 2013 guaranteed the presence of the accused’s attorney in the proceedings. The present study aims at assessing the result of the aforementioned guarantee in practice and made attempts to investigate the effect of the presence of accused’s attorney on reducing the punishment by asking the question and addressing the statistical population of this study including 48 judges of lower courts and courts of appeal. It seems that in despite of guarantees provided in the new Code of Criminal Procedure, Iran's penal system, does not tolerate the presence of an attorney in practice.Keywords: Defense attorney, equality of arms, fair trial, reducing the penalty, right to defense.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1401212 Effect of Fat Percentage and Prebiotic Composition on Proteolysis, ACE-Inhibitory and Antioxidant Activity of Probiotic Yogurt
Authors: Mohammad B. HabibiNajafi, Saeideh Sadat Fatemizadeh, Maryam Tavakoli
Abstract:
In recent years, the consumption of functional foods, including foods containing probiotic bacteria, has come to notice. Milk proteins have been identified as a source of angiotensin-I-converting enzyme )ACE( inhibitory peptides and are currently the best-known class of bioactive peptides. In this study, the effects of adding prebiotic ingredients (inulin and wheat fiber) and fat percentage (0%, 2% and 3.5%) in yogurt containing probiotic Lactobacillus casei on physicochemical properties, degree of proteolysis, antioxidant and ACE-inhibitory activity within 21 days of storage at 5 ± 1 °C were evaluated. The results of statistical analysis showed that the application of prebiotic compounds led to a significant increase in water holding capacity, proteolysis and ACE-inhibitory of samples. The degree of proteolysis in yogurt increases as storage time elapses (P < 0.05) but when proteolysis exceeds a certain threshold, this trend begins to decline. Also, during storage time, water holding capacity reduced initially but increased thereafter. Moreover, based on our findings, the survival of Lactobacillus casei in samples treated with inulin and wheat fiber increased significantly in comparison to the control sample (P < 0.05) whereas the effect of fat percentage on the survival of probiotic bacteria was not significant (P = 0.095). Furthermore, the effect of prebiotic ingredients and the presence of probiotic cultures on the antioxidant activity of samples was significant (P < 0.05).Keywords: Yogurt, proteolysis, ACE-inhibitory, antioxidant activity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1036211 Improving Topic Quality of Scripts by Using Scene Similarity Based Word Co-Occurrence
Authors: Yunseok Noh, Chang-Uk Kwak, Sun-Joong Kim, Seong-Bae Park
Abstract:
Scripts are one of the basic text resources to understand broadcasting contents. Topic modeling is the method to get the summary of the broadcasting contents from its scripts. Generally, scripts represent contents descriptively with directions and speeches, and provide scene segments that can be seen as semantic units. Therefore, a script can be topic modeled by treating a scene segment as a document. Because scene segments consist of speeches mainly, however, relatively small co-occurrences among words in the scene segments are observed. This causes inevitably the bad quality of topics by statistical learning method. To tackle this problem, we propose a method to improve topic quality with additional word co-occurrence information obtained using scene similarities. The main idea of improving topic quality is that the information that two or more texts are topically related can be useful to learn high quality of topics. In addition, more accurate topical representations lead to get information more accurate whether two texts are related or not. In this paper, we regard two scene segments are related if their topical similarity is high enough. We also consider that words are co-occurred if they are in topically related scene segments together. By iteratively inferring topics and determining semantically neighborhood scene segments, we draw a topic space represents broadcasting contents well. In the experiments, we showed the proposed method generates a higher quality of topics from Korean drama scripts than the baselines.Keywords: Broadcasting contents, generalized P´olya urn model, scripts, text similarity, topic model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1817210 Influence of Garbage Leachate on Soil Reaction,Salinity and Soil Organic Matter in East of Isfahan
Authors: Ebrahim Panahpour, Ali Gholami, Amir Hossein Davami
Abstract:
During this day a considerable amount of Leachate is produced with high amounts of organic material and nutrients needed plants. This study has done in order to scrutinize the effect of Leachate compost on the pH, EC and organic matter percentage in the form of statistical Factorial plan through randomizing block design with three main and two minor treatments and also three replications during three six month periods. Major treatments include N: Irrigation with the region-s well water as a control, I: Frequent irrigation with well water and Leachate, C: Mixing Leachate and water well (25 percent leachate + 75 percent ordinary well water) and secondary treatments, include DI: surface drip irrigation and SDI: sub surface drip irrigation. Results of this study indicated significant differences between treatments and also there were mixing up with the control treatment in the reduction of pH, increasing soluble salts and also increasing the organic matter percentage. This increase is proportional to the amount of added Leachate and in the treatment also proportional to higher mixture of frequent treatment. Therefore, since creating an acidic pH increases the ability to absorb some nutrient elements such as phosphorus, iron, zinc, copper and manganese are increased and the other hand, organic materials also improve many physical and chemical properties of soil are used in Leachate trash Consider health issues as refined in the green belts around cities as a liquid fertilizer recommended.
Keywords: Leachate, compost, drip irrigation, liquid fertilizer, soil reaction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2280209 Analysis of Temperature Change under Global Warming Impact using Empirical Mode Decomposition
Authors: Md. Khademul Islam Molla, Akimasa Sumi, M. Sayedur Rahman
Abstract:
The empirical mode decomposition (EMD) represents any time series into a finite set of basis functions. The bases are termed as intrinsic mode functions (IMFs) which are mutually orthogonal containing minimum amount of cross-information. The EMD successively extracts the IMFs with the highest local frequencies in a recursive way, which yields effectively a set low-pass filters based entirely on the properties exhibited by the data. In this paper, EMD is applied to explore the properties of the multi-year air temperature and to observe its effects on climate change under global warming. This method decomposes the original time-series into intrinsic time scale. It is capable of analyzing nonlinear, non-stationary climatic time series that cause problems to many linear statistical methods and their users. The analysis results show that the mode of EMD presents seasonal variability. The most of the IMFs have normal distribution and the energy density distribution of the IMFs satisfies Chi-square distribution. The IMFs are more effective in isolating physical processes of various time-scales and also statistically significant. The analysis results also show that the EMD method provides a good job to find many characteristics on inter annual climate. The results suggest that climate fluctuations of every single element such as temperature are the results of variations in the global atmospheric circulation.
Keywords: Empirical mode decomposition, instantaneous frequency, Hilbert spectrum, Chi-square distribution, anthropogenic impact.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2149208 Quality of Concrete of Recent Development Projects in Libya
Authors: Mohamed .S .Alazhari, Milad. M. Al Shebani
Abstract:
Numerous concrete structures projects are currently running in Libya as part of a US$50 billion government funding. The quality of concrete used in 20 different construction projects were assessed based mainly on the concrete compressive strength achieved. The projects are scattered all over the country and are at various levels of completeness. For most of these projects, the concrete compressive strength was obtained from test results of a 150mm standard cube mold. Statistical analysis of collected concrete compressive strengths reveals that the data in general followed a normal distribution pattern. The study covers comparison and assessment of concrete quality aspects such as: quality control, strength range, data standard deviation, data scatter, and ratio of minimum strength to design strength. Site quality control for these projects ranged from very good to poor according to ACI214 criteria [1]. The ranges (Rg) of the strength (max. strength – min. strength) divided by average strength are from (34% to 160%). Data scatter is measured as the range (Rg) divided by standard deviation () and is found to be (1.82 to 11.04), indicating that the range is ±3σ. International construction companies working in Libya follow different assessment criteria for concrete compressive strength in lieu of national unified procedure. The study reveals that assessments of concrete quality conducted by these construction companies usually meet their adopted (internal) standards, but sometimes fail to meet internationally known standard requirements. The assessment of concrete presented in this paper is based on ACI, British standards and proposed Libyan concrete strength assessment criteria.Keywords: Acceptance criteria, Concrete, Compressive strength, quality control
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1770