Search results for: statistical tools
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7391

Search results for: statistical tools

7271 Information Exchange Process Analysis between Authoring Design Tools and Lighting Simulation Tools

Authors: Rudan Xue, Annika Moscati, Rehel Zeleke Kebede, Peter Johansson

Abstract:

Successful buildings’ simulation and analysis inevitably require information exchange between multiple building information modeling (BIM) software. The BIM infor-mation exchange based on IFC is widely used. However, Industry Foundation Classifi-cation (IFC) files are not always reliable and information can get lost when using dif-ferent software for modeling and simulations. In this research, interviews with lighting simulation experts and a case study provided by a company producing lighting devices have been the research methods used to identify the necessary steps and data for suc-cessful information exchange between lighting simulation tools and authoring design tools. Model creation, information exchange, and model simulation have been identi-fied as key aspects for the success of information exchange. The paper concludes with recommendations for improved information exchange and more reliable simulations that take all the needed parameters into consideration.

Keywords: BIM, data exchange, interoperability issues, lighting simulations

Procedia PDF Downloads 210
7270 Hybrid Project Management Model Based on Lean and Agile Approach

Authors: Fatima-Zahra Eddoug, Jamal Benhra, Rajaa Benabbou

Abstract:

Several project management models exist in the literature and the most used ones are the hybrids for their multiple advantages. Our objective in this paper is to analyze the existing models, which are based on the Lean and Agile approaches and to propose a novel framework with the convenient tools that will allow efficient management of a general project. To create the desired framework, we were based essentially on 7 existing models. Only the Scrum tool among the agile tools was identified by several authors to be appropriate for project management. In contrast, multiple lean tools were proposed in different phases of the project.

Keywords: agility, hybrid project management, lean, scrum

Procedia PDF Downloads 113
7269 Predictive Modelling Approaches in Food Processing and Safety

Authors: Amandeep Sharma, Digvaijay Verma, Ruplal Choudhary

Abstract:

Food processing is an activity across the globe that help in better handling of agricultural produce, including dairy, meat, and fish. The operations carried out in the food industry includes raw material quality authenticity; sorting and grading; processing into various products using thermal treatments – heating, freezing, and chilling; packaging; and storage at the appropriate temperature to maximize the shelf life of the products. All this is done to safeguard the food products and to ensure the distribution up to the consumer. The approaches to develop predictive models based on mathematical or statistical tools or empirical models’ development has been reported for various milk processing activities, including plant maintenance and wastage. Recently AI is the key factor for the fourth industrial revolution. AI plays a vital role in the food industry, not only in quality and food security but also in different areas such as manufacturing, packaging, and cleaning. A new conceptual model was developed, which shows that smaller sample size as only spectra would be required to predict the other values hence leads to saving on raw materials and chemicals otherwise used for experimentation during the research and new product development activity. It would be a futuristic approach if these tools can be further clubbed with the mobile phones through some software development for their real time application in the field for quality check and traceability of the product.

Keywords: predictive modlleing, ann, ai, food

Procedia PDF Downloads 55
7268 A Review on Existing Challenges of Data Mining and Future Research Perspectives

Authors: Hema Bhardwaj, D. Srinivasa Rao

Abstract:

Technology for analysing, processing, and extracting meaningful data from enormous and complicated datasets can be termed as "big data." The technique of big data mining and big data analysis is extremely helpful for business movements such as making decisions, building organisational plans, researching the market efficiently, improving sales, etc., because typical management tools cannot handle such complicated datasets. Special computational and statistical issues, such as measurement errors, noise accumulation, spurious correlation, and storage and scalability limitations, are brought on by big data. These unique problems call for new computational and statistical paradigms. This research paper offers an overview of the literature on big data mining, its process, along with problems and difficulties, with a focus on the unique characteristics of big data. Organizations have several difficulties when undertaking data mining, which has an impact on their decision-making. Every day, terabytes of data are produced, yet only around 1% of that data is really analyzed. The idea of the mining and analysis of data and knowledge discovery techniques that have recently been created with practical application systems is presented in this study. This article's conclusion also includes a list of issues and difficulties for further research in the area. The report discusses the management's main big data and data mining challenges.

Keywords: big data, data mining, data analysis, knowledge discovery techniques, data mining challenges

Procedia PDF Downloads 88
7267 Prediction of Damage to Cutting Tools in an Earth Pressure Balance Tunnel Boring Machine EPB TBM: A Case Study L3 Guadalajara Metro Line (Mexico)

Authors: Silvia Arrate, Waldo Salud, Eloy París

Abstract:

The wear of cutting tools is one of the most decisive elements when planning tunneling works, programming the maintenance stops and saving the optimum stock of spare parts during the evolution of the excavation. Being able to predict the behavior of cutting tools can give a very competitive advantage in terms of costs and excavation performance, optimized to the needs of the TBM itself. The incredible evolution of data science in recent years gives the option to implement it at the time of analyzing the key and most critical parameters related to machinery with the purpose of knowing how the cutting head is performing in front of the excavated ground. Taking this as a case study, Metro Line 3 of Guadalajara in Mexico will develop the feasibility of using Specific Energy versus data science applied over parameters of Torque, Penetration, and Contact Force, among others, to predict the behavior and status of cutting tools. The results obtained through both techniques are analyzed and verified in the function of the wear and the field situations observed in the excavation in order to determine its effectiveness regarding its predictive capacity. In conclusion, the possibilities and improvements offered by the application of digital tools and the programming of calculation algorithms for the analysis of wear of cutting head elements compared to purely empirical methods allow early detection of possible damage to cutting tools, which is reflected in optimization of excavation performance and a significant improvement in costs and deadlines.

Keywords: cutting tools, data science, prediction, TBM, wear

Procedia PDF Downloads 26
7266 Modeling and Statistical Analysis of a Soap Production Mix in Bejoy Manufacturing Industry, Anambra State, Nigeria

Authors: Okolie Chukwulozie Paul, Iwenofu Chinwe Onyedika, Sinebe Jude Ebieladoh, M. C. Nwosu

Abstract:

The research work is based on the statistical analysis of the processing data. The essence is to analyze the data statistically and to generate a design model for the production mix of soap manufacturing products in Bejoy manufacturing company Nkpologwu, Aguata Local Government Area, Anambra state, Nigeria. The statistical analysis shows the statistical analysis and the correlation of the data. T test, Partial correlation and bi-variate correlation were used to understand what the data portrays. The design model developed was used to model the data production yield and the correlation of the variables show that the R2 is 98.7%. However, the results confirm that the data is fit for further analysis and modeling. This was proved by the correlation and the R-squared.

Keywords: General Linear Model, correlation, variables, pearson, significance, T-test, soap, production mix and statistic

Procedia PDF Downloads 418
7265 Quantitative Assessment of Soft Tissues by Statistical Analysis of Ultrasound Backscattered Signals

Authors: Da-Ming Huang, Ya-Ting Tsai, Shyh-Hau Wang

Abstract:

Ultrasound signals backscattered from the soft tissues are mainly depending on the size, density, distribution, and other elastic properties of scatterers in the interrogated sample volume. The quantitative analysis of ultrasonic backscattering is frequently implemented using the statistical approach due to that of backscattering signals tends to be with the nature of the random variable. Thus, the statistical analysis, such as Nakagami statistics, has been applied to characterize the density and distribution of scatterers of a sample. Yet, the accuracy of statistical analysis could be readily affected by the receiving signals associated with the nature of incident ultrasound wave and acoustical properties of samples. Thus, in the present study, efforts were made to explore such effects as the ultrasound operational modes and attenuation of biological tissue on the estimation of corresponding Nakagami statistical parameter (m parameter). In vitro measurements were performed from healthy and pathological fibrosis porcine livers using different single-element ultrasound transducers and duty cycles of incident tone burst ranging respectively from 3.5 to 7.5 MHz and 10 to 50%. Results demonstrated that the estimated m parameter tends to be sensitively affected by the use of ultrasound operational modes as well as the tissue attenuation. The healthy and pathological tissues may be characterized quantitatively by m parameter under fixed measurement conditions and proper calibration.

Keywords: ultrasound backscattering, statistical analysis, operational mode, attenuation

Procedia PDF Downloads 295
7264 The Effectiveness of Exchange of Tacit and Explicit Knowledge Using Digital and Face to Face Sharing

Authors: Delio I. Castaneda, Paul Toulson

Abstract:

The purpose of this study was to investigate the knowledge sharing effectiveness of two types of knowledge, tacit and explicit, depending on two channels: face to face or digital. Participants were 217 knowledge workers in New Zealand and researchers who attended a knowledge management conference in the United Kingdom. In the study, it was found that digital tools are effective to share explicit knowledge. In addition, digital tools that facilitated dialogue were effective to share tacit knowledge. It was also found that face to face communication was an effective way to share tacit and explicit knowledge. Results of this study contribute to clarify in what cases digital tools are effective to share tacit knowledge. Additionally, even though explicit knowledge can be easily shared using digital tools, this type of knowledge is also possible to be shared through dialogue. Result of this study may support practitioners to redesign programs and activities based on knowledge sharing to make strategies more effective.

Keywords: digital knowledge, explicit knowledge, knowledge sharing, tacit knowledge

Procedia PDF Downloads 232
7263 Evaluation of Free Technologies as Tools for Business Process Management

Authors: Julio Sotomayor, Daniel Yucra, Jorge Mayhuasca

Abstract:

The article presents an evaluation of free technologies for business process automation, with emphasis only on tools compatible with the general public license (GPL). The compendium of technologies was based on promoting a service-oriented enterprise architecture (SOA) and the establishment of a business process management system (BPMS). The methodology for the selection of tools was Agile UP. This proposal allows businesses to achieve technological sovereignty and independence, in addition to the promotion of service orientation and the development of free software based on components.

Keywords: BPM, BPMS suite, open-source software, SOA, enterprise architecture, business process management

Procedia PDF Downloads 260
7262 Study on Evaluating the Utilization of Social Media Tools (SMT) in Collaborative Learning Case Study: Faculty of Medicine, King Khalid University

Authors: Vasanthi Muniasamy, Intisar Magboul Ejalani, M.Anandhavalli, K. Gauthaman

Abstract:

Social Media (SM) are websites increasingly popular and built to allow people to express themselves and to interact socially with others. Most SMT are dominated by youth particularly college students. The proliferation of popular social media tools, which can accessed from any communication devices has become pervasive in the lives of today’s student life. Connecting traditional education to social media tools are a relatively new era and any collaborative tool could be used for learning activities. This study focuses (i) how the social media tools are useful for the learning activities of the students of faculty of medicine in King Khalid University (ii) whether the social media affects the collaborative learning with interaction among students, among course instructor, their engagement, perceived ease of use and perceived ease of usefulness (TAM) (iii) overall, the students satisfy with this collaborative learning through Social media.

Keywords: social media, Web 2.0, perceived ease of use, perceived usefulness, collaborative Learning

Procedia PDF Downloads 484
7261 Measurement Tools of the Maturity Model for IT Service Outsourcing in Higher Education Institutions

Authors: Victoriano Valencia García, Luis Usero Aragonés, Eugenio J. Fernández Vicente

Abstract:

Nowadays, the successful implementation of ICTs is vital for almost any kind of organization. Good governance and ICT management are essential for delivering value, managing technological risks, managing resources and performance measurement. In addition, outsourcing is a strategic IT service solution which complements IT services provided internally in organizations. This paper proposes the measurement tools of a new holistic maturity model based on standards ISO/IEC 20000 and ISO/IEC 38500, and the frameworks and best practices of ITIL and COBIT, with a specific focus on IT outsourcing. These measurement tools allow independent validation and practical application in the field of higher education, using a questionnaire, metrics tables, and continuous improvement plan tables as part of the measurement process. Guidelines and standards are proposed in the model for facilitating adaptation to universities and achieving excellence in the outsourcing of IT services.

Keywords: IT governance, IT management, IT services, outsourcing, maturity model, measurement tools

Procedia PDF Downloads 566
7260 Relationship Between Reading Comprehension and Achievement in Science Among Grade Eleven Bilingual Students in a Secondary School, Thailand

Authors: Simon Mauma Efange

Abstract:

The main aims of this research were to describe, in co-relational terms, the relationship, if any, between reading comprehension and academic achievement in science studied at the secondary level and, secondly, to find out possible trends in gender differences, such as whether boys would perform better than girls or vice versa. This research employed a quantitative design. Two kinds of instruments were employed: the Oxford Online Placement Test and the Local Assessment System Test. The Oxford Online Placement Test assesses students' English level quickly and easily. The results of these tests were subjected to statistical analysis using a special statistical software called SPSS. Statistical tools such as mean, standard deviation, percentages, frequencies, t-tests, and Pearson’s coefficient of correlation were used for the analysis of the results. Results of the t-test showed that the means are significantly different. Calculating the p-value revealed that the results were extremely statistically significant at p <.05. The value of r (Pearson correlation coefficient) was 0.2868. Although technically there is a positive correlation, the relationship between the variables is only weak (the closer the value is to zero, the weaker the relationship). However, in conclusion, calculations from the t-test using SPSS revealed that the results were statistically significant at p <.05, confirming a relationship between the two variables, and high scores in reading will give rise to slightly high scores in science. The research also revealed that having a high score in reading comprehension doesn’t necessarily mean having a high score in science or vice versa. Female subjects performed much better than male subjects in both tests, which is in line with the literature reviewed for this research.

Keywords: achievement in science, achievement in English, and bilingual students, relationship

Procedia PDF Downloads 28
7259 Space Telemetry Anomaly Detection Based On Statistical PCA Algorithm

Authors: Bassem Nassar, Wessam Hussein, Medhat Mokhtar

Abstract:

The crucial concern of satellite operations is to ensure the health and safety of satellites. The worst case in this perspective is probably the loss of a mission but the more common interruption of satellite functionality can result in compromised mission objectives. All the data acquiring from the spacecraft are known as Telemetry (TM), which contains the wealth information related to the health of all its subsystems. Each single item of information is contained in a telemetry parameter, which represents a time-variant property (i.e. a status or a measurement) to be checked. As a consequence, there is a continuous improvement of TM monitoring systems in order to reduce the time required to respond to changes in a satellite's state of health. A fast conception of the current state of the satellite is thus very important in order to respond to occurring failures. Statistical multivariate latent techniques are one of the vital learning tools that are used to tackle the aforementioned problem coherently. Information extraction from such rich data sources using advanced statistical methodologies is a challenging task due to the massive volume of data. To solve this problem, in this paper, we present a proposed unsupervised learning algorithm based on Principle Component Analysis (PCA) technique. The algorithm is particularly applied on an actual remote sensing spacecraft. Data from the Attitude Determination and Control System (ADCS) was acquired under two operation conditions: normal and faulty states. The models were built and tested under these conditions and the results shows that the algorithm could successfully differentiate between these operations conditions. Furthermore, the algorithm provides competent information in prediction as well as adding more insight and physical interpretation to the ADCS operation.

Keywords: space telemetry monitoring, multivariate analysis, PCA algorithm, space operations

Procedia PDF Downloads 395
7258 Organisational Blogging: Reviewing Its Effectiveness as an Organisational Learning Tool

Authors: Gavin J. Baxter, Mark H. Stansfield

Abstract:

This paper reviews the internal use of blogs and their potential effectiveness as organisational learning tools. Prior to and since the emergence of the concept of ‘Enterprise 2.0’ there still remains a lack of empirical evidence associated with how organisations are applying social media tools and whether they are effective towards supporting organisational learning. Surprisingly, blogs, one of the more traditional social media tools, still remains under-researched in the context of ‘Enterprise 2.0’ and organisational learning. The aim of this paper is to identify the theoretical linkage between blogs and organisational learning in addition to reviewing prior research on organisational blogging with a view towards exploring why this area remains under-researched and identifying what needs to be done to try and move the area forward. Through a review of the literature, one of the principal findings of this paper is that organisational blogs, dependent on their use, do have a mutual compatibility with the interpretivist aspect of organisational learning. This paper further advocates that further empirical work in this subject area is required to substantiate this theoretical assumption.

Keywords: Enterprise 2.0, blogs, organisational learning, social media tools

Procedia PDF Downloads 259
7257 Websites for Hypothesis Testing

Authors: Frantisek Mosna

Abstract:

E-learning has become an efficient and widespread means in process of education at all branches of human activities. Statistics is not an exception. Unfortunately the main focus in the statistics teaching is usually paid to the substitution to formulas. Suitable web-sites can simplify and automate calculation and provide more attention and time to the basic principles of statistics, mathematization of real-life situations and following interpretation of results. We introduce our own web-sites for hypothesis testing. Their didactic aspects, technical possibilities of individual tools for their creating, experience and advantages or disadvantages of them are discussed in this paper. These web-sites do not substitute common statistical software but significantly improve the teaching of the statistics at universities.

Keywords: e-learning, hypothesis testing, PHP, web-sites

Procedia PDF Downloads 397
7256 Waste Prevention and Economic Policy: Policy Tools for Increasing Resource Efficiency and Savings

Authors: Sylvia Graczka

Abstract:

Waste related environmental problems are not only exploding but are also spotlighted for capacity shortages in recycling, as China announced its ban on waste imports. According to the waste hierarchy, prevention is the primary solution for waste, and also the cheapest. Waste related environmental pollution as externality puts an ever-growing burden on communities bearing the social costs. Economic policies often claim to be pro-environment, this often appears only theoretically, or at the level of principles. There are few concrete occurrences of tools in economic policies, such as green taxes, that are truly effective in stimulating the shift towards waste reduction. The paper presents theoretical economic policy tools based on literature review, and case studies on applied economic policy tools by analyzing policy papers, strategies in force, in line with ‘polluter pays’ and ‘extended producer responsibility’ principles. The study also emphasizes the differences between the broader notion of waste reduction and that of waste minimization, parallel to the difference between resource efficiency and resource savings. It also puts the issue in the context of neoclassical environmental economics and ecological economics, to present alternatives in approach. The research concludes in identifying effective economic policy tools that support the reduction of material use, and the prevention of waste. Consumer and producer awareness of waste problems and consciousness related to their choices are inevitable to make economic policy tools work effectively.

Keywords: economic policy, producer responsibility, resource efficiency, waste prevention

Procedia PDF Downloads 128
7255 Evaluation of Technology Tools for Mathematics Instruction by Novice Elementary Teachers

Authors: Christopher J. Johnston

Abstract:

This paper presents the finding of a research study in which novice (first and second year) elementary teachers (grades Kindergarten – six) evaluated various mathematics Virtual Manipulatives, websites, and Applets (tools) for use in mathematics instruction. Participants identified the criteria they used for evaluating these types of resources and provided recommendations for or against five pre-selected tools. During the study, participants participated in three data collection activities: (1) A brief Likert-scale survey which gathered information about their attitudes toward technology use; (2) An identification of criteria for evaluating technology tools; and (3) A review of five pre-selected technology tools in light of their self-identified criteria. Data were analyzed qualitatively using four theoretical categories (codes): Software Features (41%), Mathematics (26%), Learning (22%), and Motivation (11%). These four theoretical categories were then grouped into two broad categories: Content and Instruction (Mathematics and Learning), and Surface Features (Software Features and Motivation). These combined, broad categories suggest novice teachers place roughly the same weight on pedagogical features as they do technological features. Implications for mathematics teacher educators are discussed, and suggestions for future research are provided.

Keywords: mathematics education, novice teachers, technology, virtual manipulatives

Procedia PDF Downloads 112
7254 A Cross-Gender Statistical Analysis of Tuvinian Intonation Features in Comparison With Uzbek and Azerbaijani

Authors: Daria Beziakina, Elena Bulgakova

Abstract:

The paper deals with cross-gender and cross-linguistic comparison of pitch characteristics for Tuvinian with two other Turkic languages - Uzbek and Azerbaijani, based on the results of statistical analysis of pitch parameter values and intonation patterns used by male and female speakers. The main goal of our work is to obtain the ranges of pitch parameter values typical for Tuvinian speakers for the purpose of automatic language identification. We also propose a cross-gender analysis of declarative intonation in the poorly studied Tuvinian language. The ranges of pitch parameter values were obtained by means of specially developed software that deals with the distribution of pitch values and allows us to obtain statistical language-specific pitch intervals.

Keywords: speech analysis, statistical analysis, speaker recognition, identification of person

Procedia PDF Downloads 326
7253 Understanding Consumer Behaviors by Using Neuromarketing Tools and Methods

Authors: Tabrej Khan

Abstract:

Neuromarketing can refer to the commercial application of neuroscience technologies and insights to drive business further. On the other side, consumer neuroscience can be seen as the academic use of neuroscience to better understand marketing effects on consumer behavior. Consumer Neuroscience and Neuromarketing is a multidisciplinary effort between economics, psychology, and neuroscience and information technology. Traditional methods are using survey, interviews, focus group people are overtly and consciously reporting on their experience and thoughts. The unconscious side of customer behavior is largely unmeasured in the traditional methods. Neuroscience has a potential to understand the unconscious part. Through this paper, we are going to present specific results of selected tools and methods that are used to understand consumer behaviors.

Keywords: neuromarketing, neuroscience, consumer behaviors, tools

Procedia PDF Downloads 369
7252 Experimental Investigation and Numerical Simulations of the Cylindrical Machining of a Ti-6Al-4V Tree

Authors: Mohamed Sahli, David Bassir, Thierry Barriere, Xavier Roizard

Abstract:

Predicting the behaviour of the Ti-6Al-4V alloy during the turning operation was very important in the choice of suitable cutting tools and also in the machining strategies. In this study, a 3D model with thermo-mechanical coupling has been proposed to study the influence of cutting parameters and also lubrication on the performance of cutting tools. The constants of the constitutive Johnson-Cook model of Ti-6Al-4V alloy were identified using inverse analysis based on the parameters of the orthogonal cutting process. Then, numerical simulations of the finishing machining operation were developed and experimentally validated for the cylindrical stock removal stage with the finishing cutting tool.

Keywords: titanium turning, cutting tools, FE simulation, chip

Procedia PDF Downloads 155
7251 Statistical Process Control in Manufacturing, a Case Study on an Iranian Automobile Company

Authors: M. E. Khiav, D. J. Borah, H. T. S. Santos, V. T. Faria

Abstract:

For automobile companies, it has become very important to ensure sound quality in manufacturing and assembling in order to prevent occurrence of defects and to reduce the amount of parts replacements to be done in the service centers during the warranty period. Statistical Process Control (SPC) is widely used as the tool to analyze the quality of such processes and plays a significant role in the improvement of the processes by identifying the patterns and the location of the defects. In this paper, a case study has been conducted on an Iranian automobile company. This paper performs a quality analysis of a particular component called “Internal Bearing for the Back Wheel” of a particular car model, manufactured by the company, based on the 10 million data received from its service centers located all over the country. By creating control charts including X bar–S charts and EWMA charts, it has been observed after the year 2009, the specific component underwent frequent failures and there has been a sharp dip in the average distance covered by the cars till the specific component requires replacement/maintenance. Correlation analysis was performed to find out the reasons that might have affected the quality of the specific component in all the cars produced by the company after the year 2009. Apart from manufacturing issues, some political and environmental factors have been identified to have a potential impact on the quality of the component. A maiden attempt has been made to analyze the quality issues within an Iranian automobile manufacturer; such issues often get neglected in developing countries. The paper also discusses the possibility of political scenario of Iran and the country’s environmental conditions affecting the quality of the end products, which not only strengthens the extant literature but also provides a new direction for future research.

Keywords: capability analysis, car manufacturing, statistical process control, quality control, quality tools

Procedia PDF Downloads 346
7250 Development of Sleep Quality Index Using Heart Rate

Authors: Dongjoo Kim, Chang-Sik Son, Won-Seok Kang

Abstract:

Adequate sleep affects various parts of one’s overall physical and mental life. As one of the methods in determining the appropriate amount of sleep, this research presents a heart rate based sleep quality index. In order to evaluate sleep quality using the heart rate, sleep data from 280 subjects taken over one month are used. Their sleep data are categorized by a three-part heart rate range. After categorizing, some features are extracted, and the statistical significances are verified for these features. The results show that some features of this sleep quality index model have statistical significance. Thus, this heart rate based sleep quality index may be a useful discriminator of sleep.

Keywords: sleep, sleep quality, heart rate, statistical analysis

Procedia PDF Downloads 313
7249 Features of Normative and Pathological Realizations of Sibilant Sounds for Computer-Aided Pronunciation Evaluation in Children

Authors: Zuzanna Miodonska, Michal Krecichwost, Pawel Badura

Abstract:

Sigmatism (lisping) is a speech disorder in which sibilant consonants are mispronounced. The diagnosis of this phenomenon is usually based on the auditory assessment. However, the progress in speech analysis techniques creates a possibility of developing computer-aided sigmatism diagnosis tools. The aim of the study is to statistically verify whether specific acoustic features of sibilant sounds may be related to pronunciation correctness. Such knowledge can be of great importance while implementing classifiers and designing novel tools for automatic sibilants pronunciation evaluation. The study covers analysis of various speech signal measures, including features proposed in the literature for the description of normative sibilants realization. Amplitudes and frequencies of three fricative formants (FF) are extracted based on local spectral maxima of the friction noise. Skewness, kurtosis, four normalized spectral moments (SM) and 13 mel-frequency cepstral coefficients (MFCC) with their 1st and 2nd derivatives (13 Delta and 13 Delta-Delta MFCC) are included in the analysis as well. The resulting feature vector contains 51 measures. The experiments are performed on the speech corpus containing words with selected sibilant sounds (/ʃ, ʒ/) pronounced by 60 preschool children with proper pronunciation or with natural pathologies. In total, 224 /ʃ/ segments and 191 /ʒ/ segments are employed in the study. The Mann-Whitney U test is employed for the analysis of stigmatism and normative pronunciation. Statistically, significant differences are obtained in most of the proposed features in children divided into these two groups at p < 0.05. All spectral moments and fricative formants appear to be distinctive between pathology and proper pronunciation. These metrics describe the friction noise characteristic for sibilants, which makes them particularly promising for the use in sibilants evaluation tools. Correspondences found between phoneme feature values and an expert evaluation of the pronunciation correctness encourage to involve speech analysis tools in diagnosis and therapy of sigmatism. Proposed feature extraction methods could be used in a computer-assisted stigmatism diagnosis or therapy systems.

Keywords: computer-aided pronunciation evaluation, sigmatism diagnosis, speech signal analysis, statistical verification

Procedia PDF Downloads 276
7248 Statistical Characteristics of Code Formula for Design of Concrete Structures

Authors: Inyeol Paik, Ah-Ryang Kim

Abstract:

In this research, a statistical analysis is carried out to examine the statistical properties of the formula given in the design code for concrete structures. The design formulas of the Korea highway bridge design code - the limit state design method (KHBDC) which is the current national bridge design code and the design code for concrete structures by Korea Concrete Institute (KCI) are applied for the analysis. The safety levels provided by the strength formulas of the design codes are defined based on the probabilistic and statistical theory.KHBDC is a reliability-based design code. The load and resistance factors of this code were calibrated to attain the target reliability index. It is essential to define the statistical properties for the design formulas in this calibration process. In general, the statistical characteristics of a member strength are due to the following three factors. The first is due to the difference between the material strength of the actual construction and that used in the design calculation. The second is the difference between the actual dimensions of the constructed sections and those used in design calculation. The third is the difference between the strength of the actual member and the formula simplified for the design calculation. In this paper, the statistical study is focused on the third difference. The formulas for calculating the shear strength of concrete members are presented in different ways in KHBDC and KCI. In this study, the statistical properties of design formulas were obtained through comparison with the database which comprises the experimental results from the reference publications. The test specimen was either reinforced with the shear stirrup or not. For an applied database, the bias factor was about 1.12 and the coefficient of variation was about 0.18. By applying the statistical properties of the design formula to the reliability analysis, it is shown that the resistance factors of the current design codes satisfy the target reliability indexes of both codes. Also, the minimum resistance factors of the KHBDC which is written in the material resistance factor format and KCE which is in the member resistance format are obtained and the results are presented. A further research is underway to calibrate the resistance factors of the high strength and high-performance concrete design guide.

Keywords: concrete design code, reliability analysis, resistance factor, shear strength, statistical property

Procedia PDF Downloads 295
7247 Adoption of Noise and Vibration Management Tools for Major Infrastructure Projects in Sydney, Australia

Authors: Adrian Morris, Rodney Phillips, Mattia Tabacchi

Abstract:

Minimizing construction noise and vibration impacts is a key challenge for major infrastructure projects in urban environments. Before commencing construction works, Construction Noise and Vibration Management Plan (CNVMP) and Construction Noise and Vibration Impact Statements (CNVIS) are required to be prepared and submitted to the relevant government authorities for review and approval. However, the assessment of potential impacts from work activities at pre-approval stage may be inaccurate as works methodology and scheduling are yet to be determined. In response, noise and vibration management tools have been developed to refine and supplement the CNVIS as works progress. These tools have been successfully implemented in major infrastructure projects allowing contractors to plan and assess construction works in a cost effective and timely manner. As a result, noise and vibration management tools have been incorporated into management plans and are increasingly required by regulators.

Keywords: noise management, environmental noise, infrastructure projects, construction, vibration, cost effective

Procedia PDF Downloads 101
7246 Arabic Light Word Analyser: Roles with Deep Learning Approach

Authors: Mohammed Abu Shquier

Abstract:

This paper introduces a word segmentation method using the novel BP-LSTM-CRF architecture for processing semantic output training. The objective of web morphological analysis tools is to link a formal morpho-syntactic description to a lemma, along with morpho-syntactic information, a vocalized form, a vocalized analysis with morpho-syntactic information, and a list of paradigms. A key objective is to continuously enhance the proposed system through an inductive learning approach that considers semantic influences. The system is currently under construction and development based on data-driven learning. To evaluate the tool, an experiment on homograph analysis was conducted. The tool also encompasses the assumption of deep binary segmentation hypotheses, the arbitrary choice of trigram or n-gram continuation probabilities, language limitations, and morphology for both Modern Standard Arabic (MSA) and Dialectal Arabic (DA), which provide justification for updating this system. Most Arabic word analysis systems are based on the phonotactic morpho-syntactic analysis of a word transmitted using lexical rules, which are mainly used in MENA language technology tools, without taking into account contextual or semantic morphological implications. Therefore, it is necessary to have an automatic analysis tool taking into account the word sense and not only the morpho-syntactic category. Moreover, they are also based on statistical/stochastic models. These stochastic models, such as HMMs, have shown their effectiveness in different NLP applications: part-of-speech tagging, machine translation, speech recognition, etc. As an extension, we focus on language modeling using Recurrent Neural Network (RNN); given that morphological analysis coverage was very low in dialectal Arabic, it is significantly important to investigate deeply how the dialect data influence the accuracy of these approaches by developing dialectal morphological processing tools to show that dialectal variability can support to improve analysis.

Keywords: NLP, DL, ML, analyser, MSA, RNN, CNN

Procedia PDF Downloads 18
7245 Optimization of Manufacturing Process Parameters: An Empirical Study from Taiwan's Tech Companies

Authors: Chao-Ton Su, Li-Fei Chen

Abstract:

The parameter design is crucial to improving the uniformity of a product or process. In the product design stage, parameter design aims to determine the optimal settings for the parameters of each element in the system, thereby minimizing the functional deviations of the product. In the process design stage, parameter design aims to determine the operating settings of the manufacturing processes so that non-uniformity in manufacturing processes can be minimized. The parameter design, trying to minimize the influence of noise on the manufacturing system, plays an important role in the high-tech companies. Taiwan has many well-known high-tech companies, which show key roles in the global economy. Quality remains the most important factor that enables these companies to sustain their competitive advantage. In Taiwan however, many high-tech companies face various quality problems. A common challenge is related to root causes and defect patterns. In the R&D stage, root causes are often unknown, and defect patterns are difficult to classify. Additionally, data collection is not easy. Even when high-volume data can be collected, data interpretation is difficult. To overcome these challenges, high-tech companies in Taiwan use more advanced quality improvement tools. In addition to traditional statistical methods and quality tools, the new trend is the application of powerful tools, such as neural network, fuzzy theory, data mining, industrial engineering, operations research, and innovation skills. In this study, several examples of optimizing the parameter settings for the manufacturing process in Taiwan’s tech companies will be presented to illustrate proposed approach’s effectiveness. Finally, a discussion of using traditional experimental design versus the proposed approach for process optimization will be made.

Keywords: quality engineering, parameter design, neural network, genetic algorithm, experimental design

Procedia PDF Downloads 124
7244 Cutting Tools in Finishing Operations for CNC Rapid Manufacturing Processes: Experimental Studies

Authors: M. N. Osman Zahid, K. Case, D. Watts

Abstract:

This paper reports an advanced approach in the application of CNC machining for rapid manufacturing processes (CNC-RM). The aim of this study is to improve the quality of machined parts by introducing different cutting tools during finishing operations. As the cutting is performed in different directions, the surfaces presented on part can be classified into several categories. Therefore, suitable cutting tools are assigned to machine particular surfaces and to improve the quality. Experimental studies have been carried out by fabricating several parts based on the suggested approach. The results provide further support for implementing this approach in rapid machining processes.

Keywords: CNC machining, end mill tool, finishing operation, rapid manufacturing

Procedia PDF Downloads 325
7243 Clinical Feature Analysis and Prediction on Recurrence in Cervical Cancer

Authors: Ravinder Bahl, Jamini Sharma

Abstract:

The paper demonstrates analysis of the cervical cancer based on a probabilistic model. It involves technique for classification and prediction by recognizing typical and diagnostically most important test features relating to cervical cancer. The main contributions of the research include predicting the probability of recurrences in no recurrence (first time detection) cases. The combination of the conventional statistical and machine learning tools is applied for the analysis. Experimental study with real data demonstrates the feasibility and potential of the proposed approach for the said cause.

Keywords: cervical cancer, recurrence, no recurrence, probabilistic, classification, prediction, machine learning

Procedia PDF Downloads 338
7242 Statistical Model of Water Quality in Estero El Macho, Machala-El Oro

Authors: Rafael Zhindon Almeida

Abstract:

Surface water quality is an important concern for the evaluation and prediction of water quality conditions. The objective of this study is to develop a statistical model that can accurately predict the water quality of the El Macho estuary in the city of Machala, El Oro province. The methodology employed in this study is of a basic type that involves a thorough search for theoretical foundations to improve the understanding of statistical modeling for water quality analysis. The research design is correlational, using a multivariate statistical model involving multiple linear regression and principal component analysis. The results indicate that water quality parameters such as fecal coliforms, biochemical oxygen demand, chemical oxygen demand, iron and dissolved oxygen exceed the allowable limits. The water of the El Macho estuary is determined to be below the required water quality criteria. The multiple linear regression model, based on chemical oxygen demand and total dissolved solids, explains 99.9% of the variance of the dependent variable. In addition, principal component analysis shows that the model has an explanatory power of 86.242%. The study successfully developed a statistical model to evaluate the water quality of the El Macho estuary. The estuary did not meet the water quality criteria, with several parameters exceeding the allowable limits. The multiple linear regression model and principal component analysis provide valuable information on the relationship between the various water quality parameters. The findings of the study emphasize the need for immediate action to improve the water quality of the El Macho estuary to ensure the preservation and protection of this valuable natural resource.

Keywords: statistical modeling, water quality, multiple linear regression, principal components, statistical models

Procedia PDF Downloads 62