Search results for: traditional techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11156

Search results for: traditional techniques

9236 Educational Data Mining: The Case of the Department of Mathematics and Computing in the Period 2009-2018

Authors: Mário Ernesto Sitoe, Orlando Zacarias

Abstract:

University education is influenced by several factors that range from the adoption of strategies to strengthen the whole process to the academic performance improvement of the students themselves. This work uses data mining techniques to develop a predictive model to identify students with a tendency to evasion and retention. To this end, a database of real students’ data from the Department of University Admission (DAU) and the Department of Mathematics and Informatics (DMI) was used. The data comprised 388 undergraduate students admitted in the years 2009 to 2014. The Weka tool was used for model building, using three different techniques, namely: K-nearest neighbor, random forest, and logistic regression. To allow for training on multiple train-test splits, a cross-validation approach was employed with a varying number of folds. To reduce bias variance and improve the performance of the models, ensemble methods of Bagging and Stacking were used. After comparing the results obtained by the three classifiers, Logistic Regression using Bagging with seven folds obtained the best performance, showing results above 90% in all evaluated metrics: accuracy, rate of true positives, and precision. Retention is the most common tendency.

Keywords: evasion and retention, cross-validation, bagging, stacking

Procedia PDF Downloads 85
9235 Transformational Entrepreneurship: Exploring Pedagogy in Tertiary Education

Authors: S. Karmokar

Abstract:

Over the last 20 years, there has been increasing interest in the topic of entrepreneurship education as it is seen in many countries as a way of enhancing the enterprise culture and promote capability building among community. There is also rapid growth of emerging technologies across the globe and forced entrepreneurs to searching for a new model of economic growth. There are two movements that are dominating and creating waves, Technology Entrepreneurship and Social Entrepreneurship. An increasing number of entrepreneurs are awakening to the possibility of combining the scalable tools and methodology of Technology Entrepreneurship with the value system of Social Entrepreneurship–‘Transformational Entrepreneurship’. To do this transitional educational institute’s need to figure out how to unite the scalable tools of Technology Entrepreneurship with the moral ethos of Social Entrepreneurship. The traditional entrepreneurship education model is wedded to top-down instructive approaches, that is widely used in management education have led to passive educational model. Despite the effort, disruptive’ pedagogies are rare in higher education; they remain underused and often marginalized. High impact and transformational entrepreneurship education and training require universities to adopt new practices and revise current, traditional ways of working. This is a conceptual research paper exploring the potential and growth of transformational entrepreneurship, investigating links between social entrepreneurship. Based on empirical studies and theoretical approaches, this paper outlines some educational approach for both academics and educational institutes to deliver emerging transformational entrepreneurship in tertiary education. The paper presents recommendations for tertiary educators to inform the designing of teaching practices, revise current delivery methods and encourage students to fulfill their potential as entrepreneurs.

Keywords: educational pedagogies, emerging technologies, social entrepreneurship, transformational entrepreneurship

Procedia PDF Downloads 193
9234 Improving the Strength Characteristics of Soil Using Cotton Fibers

Authors: Bindhu Lal, Karnika Kochal

Abstract:

Clayey soil contains clay minerals with traces of metal oxides and organic matter, which exhibits properties like low drainage, high plasticity, and shrinkage. To overcome these issues, various soil reinforcement techniques are used to elevate the stiffness, water tightness, and bearing capacity of the soil. Such techniques include cementation, bituminization, freezing, fiber inclusion, geo-synthetics, nailing, etc. Reinforcement of soil with fibers has been a cost-effective solution to soil improvement problems. An experimental study was undertaken involving the inclusion of cotton waste fibers in clayey soil as reinforcement with different fiber contents (1%, 1.5%, 2%, and 2.5% by weight) and analyzing its effects on the unconfined compressive strength of the soil. Two categories of soil were taken, comprising of natural clay and clay mixed with 5% sodium bentonite by weight. The soil specimens were subjected to proctor compaction and unconfined compression tests. The validated outcome shows that fiber inclusion has a strikingly positive impact on the compressive strength and axial strain at failure of the soil. Based on the commendatory results procured, compressive strength was found to be directly proportional to the fiber content, with the effect being more pronounced at lower water content.

Keywords: bentonite clay, clay, cotton fibers, unconfined compressive strength

Procedia PDF Downloads 180
9233 An Investigation into the Current Implementation of Design-Build Contracts in the Kingdom of Saudi Arabia

Authors: Ibrahim A. Alhammad, Suleiman A. Al-Otaibi, Khalid S. Al-Gahtani, Naïf Al-Otaibi, Abdulaziz A. Bubshait

Abstract:

In the last decade, the use of project delivery system of design build engineering contracts is increasing in North America due to the reasons of reducing the project duration and minimizing costs. The shift from traditional approach of Design-Bid-Build to Design-Build contracts have been attributed to many factors such as evolution of the regulatory and legal frameworks governing the engineering contracts and improvement in integrating design and construction. The aforementioned practice of contracting is more appropriate in North America; yet, it may not be the case in Saudi Arabia where the traditional approach of construction contracting remains dominant. The authors believe there are number of factors related to the gaps in the level of sophistication of the engineering and management of the construction projects in both countries. A step towards improving the Saudi construction practice by adopting the new trend of construction contracting, this paper identifies the reasons why Design/Build form of contracting are not frequently utilized. A field survey, which includes the questionnaire addressing the research problem, is distributed to three main parties of the construction contracts: clients, consultants, and contractors. The analyzed collected data were statistically sufficient to finding the reasons of not adopting the new trend of good practice of deign build approach in Saudi Arabia. In addition, the reasons are: (1) lack of regulation and legal framework; (2) absence of clear criteria of the owner for the trade-off between competing contractors, (3) and lack of experience, knowledge and skill.

Keywords: design built projects, Saudi Arabia, GCC, mega projects

Procedia PDF Downloads 221
9232 Management of ASD with Co-Morbid OCD: A Literature Review to Compare the Pharmacological and Psychological Treatment Options in Individuals Under the Age of 18

Authors: Melissa Nelson, Simran Jandu, Hana Jalal, Mia Ingram, Chrysi Stefanidou

Abstract:

There is a significant overlap between autism spectrum disorder (ASD) and obsessive compulsive disorder (OCD), with up to 90% of young people diagnosed with ASD having this co-morbidity. Distinguishing between the symptoms of the two leads to issues with accurate treatment, yet this is paramount in benefitting the young person. There are two distinct methods of treatment, psychological or pharmacological, with clinicians tending to choose one or the other, potentially due to the lack of research available. This report reviews the efficacy of psychological and pharmacological treatments for young people diagnosed with ASD and co-morbid OCD. A literature review was performed on papers from the last fifteen years, including “ASD,” “OCD,” and individuals under the age of 18. Eleven papers were selected as relevant. The report looks at the comparison between more traditional methods, such as selective serotonin reuptake inhibitors (SSRI) and cognitive behavior therapy (CBT), and newer therapies, such as modified or intensive ASD-focused psychotherapies and the use of other medication classes. On reviewing the data, it was identified that there was a distinct lack of information on this important topic. The most widely used treatment was medication such as Fluoxetine, an SSRI, which rarely showed an improvement in symptoms or outcomes. This is in contrast to modified forms of CBT, which often reduces symptoms or even results in OCD remission. With increased research into the non-traditional management of these co-morbid conditions, it is clear there is scope that modified CBT may become the future treatment of choice for OCD in young people with ASD.

Keywords: autism spectrum disorder, intensive or adapted cognitive behavioral therapy, obsessive compulsive disorder, pharmacological management

Procedia PDF Downloads 14
9231 A Review of Benefit-Risk Assessment over the Product Lifecycle

Authors: M. Miljkovic, A. Urakpo, M. Simic-Koumoutsaris

Abstract:

Benefit-risk assessment (BRA) is a valuable tool that takes place in multiple stages during a medicine's lifecycle, and this assessment can be conducted in a variety of ways. The aim was to summarize current BRA methods used during approval decisions and in post-approval settings and to see possible future directions. Relevant reviews, recommendations, and guidelines published in medical literature and through regulatory agencies over the past five years have been examined. BRA implies the review of two dimensions: the dimension of benefits (determined mainly by the therapeutic efficacy) and the dimension of risks (comprises the safety profile of a drug). Regulators, industry, and academia have developed various approaches, ranging from descriptive textual (qualitative) to decision-analytic (quantitative) models, to facilitate the BRA of medicines during the product lifecycle (from Phase I trials, to authorization procedure, post-marketing surveillance and health technology assessment for inclusion in public formularies). These approaches can be classified into the following categories: stepwise structured approaches (frameworks); measures for benefits and risks that are usually endpoint specific (metrics), simulation techniques and meta-analysis (estimation techniques), and utility survey techniques to elicit stakeholders’ preferences (utilities). All these approaches share the following two common goals: to assist this analysis and to improve the communication of decisions, but each is subject to its own specific strengths and limitations. Before using any method, its utility, complexity, the extent to which it is established, and the ease of results interpretation should be considered. Despite widespread and long-time use, BRA is subject to debate, suffers from a number of limitations, and currently is still under development. The use of formal, systematic structured approaches to BRA for regulatory decision-making and quantitative methods to support BRA during the product lifecycle is a standard practice in medicine that is subject to continuous improvement and modernization, not only in methodology but also in cooperation between organizations.

Keywords: benefit-risk assessment, benefit-risk profile, product lifecycle, quantitative methods, structured approaches

Procedia PDF Downloads 159
9230 Residential and Care Model for Elderly People Based on “Internet Plus”

Authors: Haoyi Sheng

Abstract:

China's aging tendency is becoming increasingly severe, which leads to the embarrassing situation of "getting old before getting wealthy". The traditional pension model does not comply with the need of today. Relying on "Internet Plus", it can efficiently integrate information and resources and meet the personalized needs of elderly care. It can reduce the operating cost of community elderly care facilities and lay a technical foundation for providing better services for the elderly. The key for providing help for the elderly in the future is to effectively integrate technology, make good use of technology, and improve the efficiency of elderly care services. The effective integration of traditional home care, community care, intelligent elderly care equipment and medical resources to create the "Internet Plus" community intelligent pension service mode has become the future development trend of aging care. The research method of this paper is to collect literature and conduct theoretical research on community pension firstly. Secondly, the combination of suitable aging design and "Internet Plus" is elaborated through research. Finally, this paper states the current level of intelligent technology in old-age care and looks into the future by understanding multiple levels of "Internet Plus". The development of community intelligent pension mode and content under "Internet Plus" has enormous development potential. In addition to the characteristics and functions of ordinary houses, residential design of endowment housing has higher requirements for comfort and personalization, and the people-oriented is the principle of design.

Keywords: ageing tendency, 'Internet Plus', community intelligent elderly care, elderly care service model, technology

Procedia PDF Downloads 138
9229 A Metallography Study of Secondary A226 Aluminium Alloy Used in Automotive Industries

Authors: Lenka Hurtalová, Eva Tillová, Mária Chalupová, Juraj Belan, Milan Uhríčik

Abstract:

The secondary alloy A226 is used for many automotive casting produced by mould casting and high pressure die-casting. This alloy has excellent castability, good mechanical properties and cost-effectiveness. Production of primary aluminium alloys belong to heavy source fouling of life environs. The European Union calls for the emission reduction and reduction in energy consumption, therefore, increase production of recycled (secondary) aluminium cast alloys. The contribution is deal with influence of recycling on the quality of the casting made from A226 in automotive industry. The properties of the casting made from secondary aluminium alloys were compared with the required properties of primary aluminium alloys. The effect of recycling on microstructure was observed using combination different analytical techniques (light microscopy upon black-white etching, scanning electron microscopy-SEM upon deep etching and energy dispersive X-ray analysis-EDX). These techniques were used for the identification of the various structure parameters, which was used to compare secondary alloy microstructure with primary alloy microstructure.

Keywords: A226 secondary aluminium alloy, deep etching, mechanical properties, recycling foundry aluminium alloy

Procedia PDF Downloads 544
9228 Geovisualization of Human Mobility Patterns in Los Angeles Using Twitter Data

Authors: Linna Li

Abstract:

The capability to move around places is doubtless very important for individuals to maintain good health and social functions. People’s activities in space and time have long been a research topic in behavioral and socio-economic studies, particularly focusing on the highly dynamic urban environment. By analyzing groups of people who share similar activity patterns, many socio-economic and socio-demographic problems and their relationships with individual behavior preferences can be revealed. Los Angeles, known for its large population, ethnic diversity, cultural mixing, and entertainment industry, faces great transportation challenges such as traffic congestion, parking difficulties, and long commuting. Understanding people’s travel behavior and movement patterns in this metropolis sheds light on potential solutions to complex problems regarding urban mobility. This project visualizes people’s trajectories in Greater Los Angeles (L.A.) Area over a period of two months using Twitter data. A Python script was used to collect georeferenced tweets within the Greater L.A. Area including Ventura, San Bernardino, Riverside, Los Angeles, and Orange counties. Information associated with tweets includes text, time, location, and user ID. Information associated with users includes name, the number of followers, etc. Both aggregated and individual activity patterns are demonstrated using various geovisualization techniques. Locations of individual Twitter users were aggregated to create a surface of activity hot spots at different time instants using kernel density estimation, which shows the dynamic flow of people’s movement throughout the metropolis in a twenty-four-hour cycle. In the 3D geovisualization interface, the z-axis indicates time that covers 24 hours, and the x-y plane shows the geographic space of the city. Any two points on the z axis can be selected for displaying activity density surface within a particular time period. In addition, daily trajectories of Twitter users were created using space-time paths that show the continuous movement of individuals throughout the day. When a personal trajectory is overlaid on top of ancillary layers including land use and road networks in 3D visualization, the vivid representation of a realistic view of the urban environment boosts situational awareness of the map reader. A comparison of the same individual’s paths on different days shows some regular patterns on weekdays for some Twitter users, but for some other users, their daily trajectories are more irregular and sporadic. This research makes contributions in two major areas: geovisualization of spatial footprints to understand travel behavior using the big data approach and dynamic representation of activity space in the Greater Los Angeles Area. Unlike traditional travel surveys, social media (e.g., Twitter) provides an inexpensive way of data collection on spatio-temporal footprints. The visualization techniques used in this project are also valuable for analyzing other spatio-temporal data in the exploratory stage, thus leading to informed decisions about generating and testing hypotheses for further investigation. The next step of this research is to separate users into different groups based on gender/ethnic origin and compare their daily trajectory patterns.

Keywords: geovisualization, human mobility pattern, Los Angeles, social media

Procedia PDF Downloads 121
9227 Introduction to Multi-Agent Deep Deterministic Policy Gradient

Authors: Xu Jie

Abstract:

As a key network security method, cryptographic services must fully cope with problems such as the wide variety of cryptographic algorithms, high concurrency requirements, random job crossovers, and instantaneous surges in workloads. Its complexity and dynamics also make it difficult for traditional static security policies to cope with the ever-changing situation. Cyber Threats and Environment. Traditional resource scheduling algorithms are inadequate when facing complex decisionmaking problems in dynamic environments. A network cryptographic resource allocation algorithm based on reinforcement learning is proposed, aiming to optimize task energy consumption, migration cost, and fitness of differentiated services (including user, data, and task security). By modeling the multi-job collaborative cryptographic service scheduling problem as a multiobjective optimized job flow scheduling problem, and using a multi-agent reinforcement learning method, efficient scheduling and optimal configuration of cryptographic service resources are achieved. By introducing reinforcement learning, resource allocation strategies can be adjusted in real time in a dynamic environment, improving resource utilization and achieving load balancing. Experimental results show that this algorithm has significant advantages in path planning length, system delay and network load balancing, and effectively solves the problem of complex resource scheduling in cryptographic services.

Keywords: multi-agent reinforcement learning, non-stationary dynamics, multi-agent systems, cooperative and competitive agents

Procedia PDF Downloads 26
9226 Reducing Crash Risk at Intersections with Safety Improvements

Authors: Upal Barua

Abstract:

Crash risk at intersections is a critical safety issue. This paper examines the effectiveness of removing an existing off-set at an intersection by realignment, in reducing crashes. Empirical Bayes method was applied to conduct a before-and-after study to assess the effect of this safety improvement. The Transportation Safety Improvement Program in Austin Transportation Department completed several safety improvement projects at high crash intersections with a view to reducing crashes. One of the common safety improvement techniques applied was the realignment of intersection approaches removing an existing off-set. This paper illustrates how this safety improvement technique is applied at a high crash intersection from inception to completion. This paper also highlights the significant crash reductions achieved from this safety improvement technique applying Empirical Bayes method in a before-and-after study. The result showed that realignment of intersection approaches removing an existing off-set can reduce crashes by 53%. This paper also features the state of the art techniques applied in planning, engineering, designing and construction of this safety improvement, key factors driving the success, and lessons learned in the process.

Keywords: crash risk, intersection, off-set, safety improvement technique, before-and-after study, empirical Bayes method

Procedia PDF Downloads 245
9225 Electrochemical and Theoretical Quantum Approaches on the Inhibition of C1018 Carbon Steel Corrosion in Acidic Medium Containing Chloride Using Newly Synthesized Phenolic Schiff Bases Compounds

Authors: Hany M. Abd El-Lateef

Abstract:

Two novel Schiff bases, 5-bromo-2-[(E)-(pyridin-3-ylimino) methyl] phenol (HBSAP) and 5-bromo-2-[(E)-(quinolin-8-ylimino) methyl] phenol (HBSAQ) have been synthesized. They have been characterized by elemental analysis and spectroscopic techniques (UV–Vis, IR and NMR). Moreover, the molecular structure of HBSAP and HBSAQ compounds are determined by single crystal X-ray diffraction technique. The inhibition activity of HBSAP and HBSAQ for carbon steel in 3.5 %NaCl+0.1 M HCl for both short and long immersion time, at different temperatures (20-50 ºC), was investigated using electrochemistry and surface characterization. The potentiodynamic polarization shows that the inhibitors molecule is more adsorbed on the cathodic sites. Its efficiency increases with increasing inhibitor concentrations (92.8 % at the optimal concentration of 10-3 M for HBSAQ). Adsorption of the inhibitors on the carbon steel surface was found to obey Langmuir’s adsorption isotherm with physical/chemical nature of the adsorption, as it is shown also by scanning electron microscopy. Further, the electronic structural calculations using quantum chemical methods were found to be in a good agreement with the results of the experimental studies.

Keywords: carbon steel, Schiff bases, corrosion inhibition, SEM, electrochemical techniques

Procedia PDF Downloads 392
9224 Methods for Enhancing Ensemble Learning or Improving Classifiers of This Technique in the Analysis and Classification of Brain Signals

Authors: Seyed Mehdi Ghezi, Hesam Hasanpoor

Abstract:

This scientific article explores enhancement methods for ensemble learning with the aim of improving the performance of classifiers in the analysis and classification of brain signals. The research approach in this field consists of two main parts, each with its own strengths and weaknesses. The choice of approach depends on the specific research question and available resources. By combining these approaches and leveraging their respective strengths, researchers can enhance the accuracy and reliability of classification results, consequently advancing our understanding of the brain and its functions. The first approach focuses on utilizing machine learning methods to identify the best features among the vast array of features present in brain signals. The selection of features varies depending on the research objective, and different techniques have been employed for this purpose. For instance, the genetic algorithm has been used in some studies to identify the best features, while optimization methods have been utilized in others to identify the most influential features. Additionally, machine learning techniques have been applied to determine the influential electrodes in classification. Ensemble learning plays a crucial role in identifying the best features that contribute to learning, thereby improving the overall results. The second approach concentrates on designing and implementing methods for selecting the best classifier or utilizing meta-classifiers to enhance the final results in ensemble learning. In a different section of the research, a single classifier is used instead of multiple classifiers, employing different sets of features to improve the results. The article provides an in-depth examination of each technique, highlighting their advantages and limitations. By integrating these techniques, researchers can enhance the performance of classifiers in the analysis and classification of brain signals. This advancement in ensemble learning methodologies contributes to a better understanding of the brain and its functions, ultimately leading to improved accuracy and reliability in brain signal analysis and classification.

Keywords: ensemble learning, brain signals, classification, feature selection, machine learning, genetic algorithm, optimization methods, influential features, influential electrodes, meta-classifiers

Procedia PDF Downloads 76
9223 Investigation of Resilient Circles in Local Community and Industry: Waju-Traditional Culture in Japan and Modern Technology Application

Authors: R. Ueda

Abstract:

Today global society is seeking resilient partnership in local organizations and individuals, which realizes multi-stakeholders relationship. Although it is proposed by modern global framework of sustainable development, it is conceivable that such affiliation can be found out in the traditional local community in Japan, and that traditional spirit is tacitly sustaining in modern context of disaster mitigation in society and economy. Then this research is aiming to clarify and analyze implication for the global world by actual case studies. Regional and urban resilience is the ability of multi-stakeholders to cooperate flexibly and to adapt in response to changes in the circumstances caused by disasters, but there are various conflicts affecting coordination of disaster relief measures. These conflicts arise not only from a lack of communication and an insufficient network, but also from the difficulty to jointly draw common context from fragmented information. This is because of the weakness of our modern engineering which focuses on maintenance and restoration of individual systems. Here local ‘circles’ holistically includes local community and interacts periodically. Focusing on examples of resilient organizations and wisdom created in communities, what can be seen throughout history is a virtuous cycle where the information and the knowledge are structured, the context to be adapted becomes clear, and an adaptation at a higher level is made possible, by which the collaboration between organizations is deepened and expanded. And the wisdom of a solid and autonomous disaster prevention formed by the historical community called’ Waju’ – an area surrounded by circle embankment to protect the settlement from flood – lives on in government efforts of the coastal industrial island of today. Industrial company there collaborates to create a circle including common evacuation space, road access improvement and infrastructure recovery. These days, people here adopts new interface technology. Large-scale AR- Augmented Reality for more than hundred people is expressing detailed hazard by tsunami and liquefaction. Common experiences of the major disaster space and circle of mutual discussion are enforcing resilience. Collaboration spirit lies in the center of circle. A consistent key point is a virtuous cycle where the information and the knowledge are structured, the context to be adapted becomes clear, and an adaptation at a higher level is made possible, by which the collaboration between organizations is deepened and expanded. This writer believes that both self-governing human organizations and the societal implementation of technical systems are necessary. Infrastructure should be autonomously instituted by associations of companies and other entities in industrial areas for working closely with local governments. To develop advanced disaster prevention and multi-stakeholder collaboration, partnerships among industry, government, academia and citizens are important.

Keywords: industrial recovery, multi-sakeholders, traditional culture, user experience, Waju

Procedia PDF Downloads 114
9222 An Argument for Agile, Lean, and Hybrid Project Management in Museum Conservation Practice: A Qualitative Evaluation of the Morris Collection Conservation Project at the Sainsbury Centre for Visual Arts

Authors: Maria Ledinskaya

Abstract:

This paper is part case study and part literature review. It seeks to introduce Agile, Lean, and Hybrid project management concepts from business, software development, and manufacturing fields to museum conservation by looking at their practical application on a recent conservation project at the Sainsbury Centre for Visual Arts. The author outlines the advantages of leaner and more agile conservation practices in today’s faster, less certain, and more budget-conscious museum climate where traditional project structures are no longer as relevant or effective. The Morris Collection Conservation Project was carried out in 2019-2021 in Norwich, UK, and concerned the remedial conservation of around 150 Abstract Constructivist artworks bequeathed to the Sainsbury Centre by private collectors Michael and Joyce Morris. It was a medium-sized conservation project of moderate complexity, planned and delivered in an environment with multiple known unknowns – unresearched collection, unknown conditions and materials, unconfirmed budget. The project was later impacted by the COVID-19 pandemic, introducing indeterminate lockdowns, budget cuts, staff changes, and the need to accommodate social distancing and remote communications. The author, then a staff conservator at the Sainsbury Centre who acted as project manager on the Morris Project, presents an incremental, iterative, and value-based approach to managing a conservation project in an uncertain environment. The paper examines the project from the point of view of Traditional, Agile, Lean, and Hybrid project management. The author argues that most academic writing on project management in conservation has focussed on a Traditional plan-driven approach – also known as Waterfall project management – which has significant drawbacks in today’s museum environment due to its over-reliance on prediction-based planning and its low tolerance to change. In the last 20 years, alternative Agile, Lean and Hybrid approaches to project management have been widely adopted in software development, manufacturing, and other industries, although their recognition in the museum sector has been slow. Using examples from the Morris Project, the author introduces key principles and tools of Agile, Lean, and Hybrid project management and presents a series of arguments on the effectiveness of these alternative methodologies in museum conservation, including the ethical and practical challenges to their implementation. These project management approaches are discussed in the context of consequentialist, relativist, and utilitarian developments in contemporary conservation ethics. Although not intentionally planned as such, the Morris Project had a number of Agile and Lean features which were instrumental to its successful delivery. These key features are identified as distributed decision-making, a co-located cross-disciplinary team, servant leadership, focus on value-added work, flexible planning done in shorter sprint cycles, light documentation, and emphasis on reducing procedural, financial, and logistical waste. Overall, the author’s findings point in favour of a hybrid model, which combines traditional and alternative project processes and tools to suit the specific needs of the project.

Keywords: agile project management, conservation, hybrid project management, lean project management, waterfall project management

Procedia PDF Downloads 71
9221 Laboratory Model Tests on Encased Group Columns

Authors: Kausar Ali

Abstract:

There are several ground treatment techniques which may meet the twin objectives of increasing the bearing capacity with simultaneous reduction of settlements, but the use of stone columns is one of the most suited techniques for flexible structures such as embankments, oil storage tanks etc. that can tolerate some settlement and used worldwide. However, when the stone columns in very soft soils are loaded; stone columns undergo excessive settlement due to low lateral confinement provided by the soft soil, leading to the failure of the structure. The poor performance of stone columns under these conditions can be improved by encasing the columns with a suitable geosynthetic. In this study, the effect of reinforcement on bearing capacity of composite soil has been investigated by conducting laboratory model tests on floating and end bearing long stone columns with l/d ratio of 12. The columns were reinforced by providing geosynthetic encasement over varying column length (upper 25%, 50%, 75%, and 100% column length). In this study, a group of columns has been used instead of single column, because in the field, columns used for the purpose always remain in groups. The tests indicate that the encasement over the full column length gives higher failure stress as compared to the encasement over the partial column length for both floating and end bearing long columns. The performance of end-bearing columns was found much better than the floating columns.

Keywords: geosynthetic, ground improvement, soft clay, stone column

Procedia PDF Downloads 433
9220 Digital Employment of Disabled People: Empirical Study from Shanghai

Authors: Yan Zi, Han Xiao

Abstract:

Across the globe, ICTs are influencing employment both as an industry that creates jobs and as a tool that empowers disabled people to access new forms of work, in innovative and more flexible ways. The advancements in ICT and the number of apps and solutions that support persons with physical, cognitive and intellectual disabilities challenge traditional biased notions and offer a pathway out of traditional sheltered workshops. As the global leader in digital technology innovation, China is arguably a leader in the use of digital technology as a 'lever' in ending the economic and social marginalization of the disabled. This study investigates factors that influence adoption and use of employment-oriented ICT applications among disabled people in China and seeks to integrate three theoretical approaches: the technology acceptance model (TAM), the uses and gratifications (U&G) approach, and the social model of disability. To that end, the study used data from self-reported survey of 214 disabled adults who have been involved in two top-down 'Internet + employment' programs promoted by local disabled persons’ federation in Shanghai. A structural equation model employed in the study demonstrates that the use of employment-oriented ICT applications is affected by demographic factors of gender, categories of disability, education and marital status. The organizational support of local social organizations demonstrates significate effects on the motivations of disabled people. Results from the focus group interviews particularly suggested that to maximize the positive impact of ICTs on employment, there is significant need to build stakeholder capacity on how ICTs could benefits persons with disabilities.

Keywords: disabled people, ICTs, technology acceptance model, uses and gratifications, the social model of disability

Procedia PDF Downloads 108
9219 Single-Molecule Analysis of Structure and Dynamics in Polymer Materials by Super-Resolution Technique

Authors: Hiroyuki Aoki

Abstract:

The physical properties of polymer materials are dependent on the conformation and molecular motion of a polymer chain. Therefore, the structure and dynamic behavior of the single polymer chain have been the most important concerns in the field of polymer physics. However, it has been impossible to directly observe the conformation of the single polymer chain in a bulk medium. In the current work, the novel techniques to study the conformation and dynamics of a single polymer chain are proposed. Since a fluorescence method is extremely sensitive, the fluorescence microscopy enables the direct detection of a single molecule. However, the structure of the polymer chain as large as 100 nm cannot be resolved by conventional fluorescence methods because of the diffraction limit of light. In order to observe the single chains, we developed the labeling method of polymer materials with a photo-switchable dye and the super-resolution microscopy. The real-space conformational analysis of single polymer chains with the spatial resolution of 15-20 nm was achieved. The super-resolution microscopy enables us to obtain the three-dimensional coordinates; therefore, we succeeded the conformational analysis in three dimensions. The direct observation by the nanometric optical microscopy would reveal the detailed information on the molecular processes in the various polymer systems.

Keywords: polymer materials, single molecule, super-resolution techniques, conformation

Procedia PDF Downloads 306
9218 Imp_hist-Si: Improved Hybrid Image Segmentation Technique for Satellite Imagery to Decrease the Segmentation Error Rate

Authors: Neetu Manocha

Abstract:

Image segmentation is a technique where a picture is parted into distinct parts having similar features which have a place with similar items. Various segmentation strategies have been proposed as of late by prominent analysts. But, after ultimate thorough research, the novelists have analyzed that generally, the old methods do not decrease the segmentation error rate. Then author finds the technique HIST-SI to decrease the segmentation error rates. In this technique, cluster-based and threshold-based segmentation techniques are merged together. After then, to improve the result of HIST-SI, the authors added the method of filtering and linking in this technique named Imp_HIST-SI to decrease the segmentation error rates. The goal of this research is to find a new technique to decrease the segmentation error rates and produce much better results than the HIST-SI technique. For testing the proposed technique, a dataset of Bhuvan – a National Geoportal developed and hosted by ISRO (Indian Space Research Organisation) is used. Experiments are conducted using Scikit-image & OpenCV tools of Python, and performance is evaluated and compared over various existing image segmentation techniques for several matrices, i.e., Mean Square Error (MSE) and Peak Signal Noise Ratio (PSNR).

Keywords: satellite image, image segmentation, edge detection, error rate, MSE, PSNR, HIST-SI, linking, filtering, imp_HIST-SI

Procedia PDF Downloads 141
9217 Impact of the Quality of Aggregate on the Elasticity Modulus of Concrete

Authors: K. Krizova

Abstract:

This objective of this article is to present concrete that differs by the size of the aggregate used. The set of concrete contained six concrete recipes manufactured as traditional vibrated concrete containing identical basic components of concrete. The experiment focused on monitoring the resulting properties of hardened concrete, specifically the primary strength and modulus of the concrete elasticity and the developing parameters from 7 to 180 days were assessed.

Keywords: aggregate, cement, concrete, elasticity modulus

Procedia PDF Downloads 318
9216 Identifying the Structural Components of Old Buildings from Floor Plans

Authors: Shi-Yu Xu

Abstract:

The top three risk factors that have contributed to building collapses during past earthquake events in Taiwan are: "irregular floor plans or elevations," "insufficient columns in single-bay buildings," and the "weak-story problem." Fortunately, these unsound structural characteristics can be directly identified from the floor plans. However, due to the vast number of old buildings, conducting manual inspections to identify these compromised structural features in all existing structures would be time-consuming and prone to human errors. This study aims to develop an algorithm that utilizes artificial intelligence techniques to automatically pinpoint the structural components within a building's floor plans. The obtained spatial information will be utilized to construct a digital structural model of the building. This information, particularly regarding the distribution of columns in the floor plan, can then be used to conduct preliminary seismic assessments of the building. The study employs various image processing and pattern recognition techniques to enhance detection efficiency and accuracy. The study enables a large-scale evaluation of structural vulnerability for numerous old buildings, providing ample time to arrange for structural retrofitting in those buildings that are at risk of significant damage or collapse during earthquakes.

Keywords: structural vulnerability detection, object recognition, seismic capacity assessment, old buildings, artificial intelligence

Procedia PDF Downloads 90
9215 A Review on Existing Challenges of Data Mining and Future Research Perspectives

Authors: Hema Bhardwaj, D. Srinivasa Rao

Abstract:

Technology for analysing, processing, and extracting meaningful data from enormous and complicated datasets can be termed as "big data." The technique of big data mining and big data analysis is extremely helpful for business movements such as making decisions, building organisational plans, researching the market efficiently, improving sales, etc., because typical management tools cannot handle such complicated datasets. Special computational and statistical issues, such as measurement errors, noise accumulation, spurious correlation, and storage and scalability limitations, are brought on by big data. These unique problems call for new computational and statistical paradigms. This research paper offers an overview of the literature on big data mining, its process, along with problems and difficulties, with a focus on the unique characteristics of big data. Organizations have several difficulties when undertaking data mining, which has an impact on their decision-making. Every day, terabytes of data are produced, yet only around 1% of that data is really analyzed. The idea of the mining and analysis of data and knowledge discovery techniques that have recently been created with practical application systems is presented in this study. This article's conclusion also includes a list of issues and difficulties for further research in the area. The report discusses the management's main big data and data mining challenges.

Keywords: big data, data mining, data analysis, knowledge discovery techniques, data mining challenges

Procedia PDF Downloads 110
9214 Performance Evaluation of Production Schedules Based on Process Mining

Authors: Kwan Hee Han

Abstract:

External environment of enterprise is rapidly changing majorly by global competition, cost reduction pressures, and new technology. In these situations, production scheduling function plays a critical role to meet customer requirements and to attain the goal of operational efficiency. It deals with short-term decision making in the production process of the whole supply chain. The major task of production scheduling is to seek a balance between customer orders and limited resources. In manufacturing companies, this task is so difficult because it should efficiently utilize resource capacity under the careful consideration of many interacting constraints. At present, many computerized software solutions have been utilized in many enterprises to generate a realistic production schedule to overcome the complexity of schedule generation. However, most production scheduling systems do not provide sufficient information about the validity of the generated schedule except limited statistics. Process mining only recently emerged as a sub-discipline of both data mining and business process management. Process mining techniques enable the useful analysis of a wide variety of processes such as process discovery, conformance checking, and bottleneck analysis. In this study, the performance of generated production schedule is evaluated by mining event log data of production scheduling software system by using the process mining techniques since every software system generates event logs for the further use such as security investigation, auditing and error bugging. An application of process mining approach is proposed for the validation of the goodness of production schedule generated by scheduling software systems in this study. By using process mining techniques, major evaluation criteria such as utilization of workstation, existence of bottleneck workstations, critical process route patterns, and work load balance of each machine over time are measured, and finally, the goodness of production schedule is evaluated. By using the proposed process mining approach for evaluating the performance of generated production schedule, the quality of production schedule of manufacturing enterprises can be improved.

Keywords: data mining, event log, process mining, production scheduling

Procedia PDF Downloads 280
9213 Effectiveness of Traditional Chinese Medicine in the Treatment of Eczema: A Systematic Review and Meta-Analysis Based on Eczema Area and Severity Index Score

Authors: Oliver Chunho Ma, Tszying Chang

Abstract:

Background: Traditional Chinese Medicine (TCM) has been widely used in the treatment of eczema. However, there is currently a lack of comprehensive research on the overall effectiveness of TCM in treating eczema, particularly using the Eczema Area and Severity Index (EASI) score as an evaluation tool. Meta-analysis can integrate the results of multiple studies to provide more convincing evidence. Objective: To conduct a systematic review and meta-analysis based on the EASI score to evaluate the overall effectiveness of TCM in the treatment of eczema. Specifically, the study will review and analyze published clinical studies that investigate TCM treatments for eczema and use the EASI score as an outcome measure, comparing the differences in improving the severity of eczema between TCM and other treatment modalities, such as conventional Western medicine treatments. Methods: Relevant studies, including randomized controlled trials (RCTs) and non-randomized controlled trials, that involve TCM treatment for eczema and use the EASI score as an outcome measure will be searched in medical literature databases such as PubMed, CNKI, etc. Relevant data will be extracted from the selected studies, including study design, sample size, treatment methods, improvement in EASI score, etc. The methodological quality and risk of bias of the included studies will be assessed using appropriate evaluation tools (such as the Cochrane Handbook). The results of the selected studies will be statistically analyzed, including pooling effect sizes (such as standardized mean differences, relative risks, etc.), subgroup analysis (e.g., different TCM syndromes, different treatment modalities), and sensitivity analysis (e.g., excluding low-quality studies). Based on the results of the statistical analysis and quality assessment, the overall effectiveness of TCM in improving the severity of eczema will be interpreted. Expected outcomes: By integrating the results of multiple studies, we expect to provide more convincing evidence regarding the specific effects of TCM in improving the severity of eczema. Additionally, subgroup analysis and sensitivity analysis can further elucidate whether the effectiveness of TCM treatment is influenced by different factors. Besides, we will compare the results of the meta-analysis with the clinical data from our clinic. For both the clinical data and the meta-analysis results, we will perform descriptive statistics such as means, standard deviations, percentages, etc. and compare the differences between the two using statistical tests such as independent samples t-test or non-parametric tests to assess the statistical differences between them.

Keywords: Eczema, traditional Chinese medicine, EASI, systematic review, meta-analysis

Procedia PDF Downloads 59
9212 Anti-Allergic Activities of Smilax Glabra Rhizome Extracts and Its Isolated Compounds

Authors: Arunporn Itharat, Kamonmas Srikwan, Srisopa Ruangnoo, Pakakrong Thongdeeying

Abstract:

Background: The rhizomes of Smilax glabra (SG) has long been used in Traditional Chinese and Thai herbal medicine to treat a variety of infectious diseases and immunological disorders. Objective: To investigate the in vitro anti-allergic activities of crude extracts and pure isolated flavonoid compounds from SG by determination of inhibitory effects on antigen-induced release of β-hexosaminidase from RBL-2H3 cells. Methods: The in vitro inhibitory effects of crude aqueous and organic extracts on beta-hexosaminidase release in RBL-2H3 cells were evaluated as an in vitro indication of possible anti-allergic activity in vivo. Bioassay-guided fractionation of extracts was used to isolate flavonoid compounds from the ethanolic extracts. Results: The 95% and 50% ethanolic extracts of SG showed remarkably high anti-allergic activity, with IC50 values of 5.74 ± 2.44 and 23.54 ± 4.75 μg/ml, much higher activity than that for Ketotifen (IC50 58.90 μM). The water extract had negligible activity (IC50 > 100 μg/ml). The two isolated flavonols, Engeletin and Astilbin, showed weak anti-allergic activity, IC50 values 97.46 ± 2.04 and > 100 μg/ml, respectively. Conclusions: The 95% and 50% ethanolic extracts of SG showed strong anti-allergic activity but two flavonol constituents did not show any significant anti-allergic activity. These findings suggest that a combination of effects of various phytochemicals in crude extracts used in traditional medicine are responsible for the purported anti-allergic activity of SG herbal preparations. The plethora of constituents in crude extracts, as yet unidentified, are likely to be acting synergistically to account for the strong observed anti-allergic in vitro activity.

Keywords: Smilax glabra, anti-allergic activity, RBL-2H3 cells, flavonoid compounds

Procedia PDF Downloads 667
9211 Normal and Peaberry Coffee Beans Classification from Green Coffee Bean Images Using Convolutional Neural Networks and Support Vector Machine

Authors: Hira Lal Gope, Hidekazu Fukai

Abstract:

The aim of this study is to develop a system which can identify and sort peaberries automatically at low cost for coffee producers in developing countries. In this paper, the focus is on the classification of peaberries and normal coffee beans using image processing and machine learning techniques. The peaberry is not bad and not a normal bean. The peaberry is born in an only single seed, relatively round seed from a coffee cherry instead of the usual flat-sided pair of beans. It has another value and flavor. To make the taste of the coffee better, it is necessary to separate the peaberry and normal bean before green coffee beans roasting. Otherwise, the taste of total beans will be mixed, and it will be bad. In roaster procedure time, all the beans shape, size, and weight must be unique; otherwise, the larger bean will take more time for roasting inside. The peaberry has a different size and different shape even though they have the same weight as normal beans. The peaberry roasts slower than other normal beans. Therefore, neither technique provides a good option to select the peaberries. Defect beans, e.g., sour, broken, black, and fade bean, are easy to check and pick up manually by hand. On the other hand, the peaberry pick up is very difficult even for trained specialists because the shape and color of the peaberry are similar to normal beans. In this study, we use image processing and machine learning techniques to discriminate the normal and peaberry bean as a part of the sorting system. As the first step, we applied Deep Convolutional Neural Networks (CNN) and Support Vector Machine (SVM) as machine learning techniques to discriminate the peaberry and normal bean. As a result, better performance was obtained with CNN than with SVM for the discrimination of the peaberry. The trained artificial neural network with high performance CPU and GPU in this work will be simply installed into the inexpensive and low in calculation Raspberry Pi system. We assume that this system will be used in under developed countries. The study evaluates and compares the feasibility of the methods in terms of accuracy of classification and processing speed.

Keywords: convolutional neural networks, coffee bean, peaberry, sorting, support vector machine

Procedia PDF Downloads 145
9210 Improved Traveling Wave Method Based Fault Location Algorithm for Multi-Terminal Transmission System of Wind Farm with Grounding Transformer

Authors: Ke Zhang, Yongli Zhu

Abstract:

Due to rapid load growths in today’s highly electrified societies and the requirement for green energy sources, large-scale wind farm power transmission system is constantly developing. This system is a typical multi-terminal power supply system, whose structure of the network topology of transmission lines is complex. What’s more, it locates in the complex terrain of mountains and grasslands, thus increasing the possibility of transmission line faults and finding the fault location with difficulty after the faults and resulting in an extremely serious phenomenon of abandoning the wind. In order to solve these problems, a fault location method for multi-terminal transmission line based on wind farm characteristics and improved single-ended traveling wave positioning method is proposed. Through studying the zero sequence current characteristics by using the characteristics of the grounding transformer(GT) in the existing large-scale wind farms, it is obtained that the criterion for judging the fault interval of the multi-terminal transmission line. When a ground short-circuit fault occurs, there is only zero sequence current on the path between GT and the fault point. Therefore, the interval where the fault point exists is obtained by determining the path of the zero sequence current. After determining the fault interval, The location of the short-circuit fault point is calculated by the traveling wave method. However, this article uses an improved traveling wave method. It makes the positioning accuracy more accurate by combining the single-ended traveling wave method with double-ended electrical data. What’s more, a method of calculating the traveling wave velocity is deduced according to the above improvements (it is the actual wave velocity in theory). The improvement of the traveling wave velocity calculation method further improves the positioning accuracy. Compared with the traditional positioning method, the average positioning error of this method is reduced by 30%.This method overcomes the shortcomings of the traditional method in poor fault location of wind farm transmission lines. In addition, it is more accurate than the traditional fixed wave velocity method in the calculation of the traveling wave velocity. It can calculate the wave velocity in real time according to the scene and solve the traveling wave velocity can’t be updated with the environment and real-time update. The method is verified in PSCAD/EMTDC.

Keywords: grounding transformer, multi-terminal transmission line, short circuit fault location, traveling wave velocity, wind farm

Procedia PDF Downloads 264
9209 Study of Two MPPTs for Photovoltaic Systems Using Controllers Based in Fuzzy Logic and Sliding Mode

Authors: N. Ould cherchali, M. S. Boucherit, L. Barazane, A. Morsli

Abstract:

Photovoltaic power is widely used to supply isolated or unpopulated areas (lighting, pumping, etc.). Great advantage is that this source is inexhaustible, it offers great safety in use and it is clean. But the dynamic models used to describe a photovoltaic system are complicated and nonlinear and due to nonlinear I-V and P–V characteristics of photovoltaic generators, a maximum power point tracking technique (MPPT) is required to maximize the output power. In this paper, two online techniques of maximum power point tracking using robust controller for photovoltaic systems are proposed, the first technique use fuzzy logic controller (FLC) and the second use sliding mode controller (SMC) for photovoltaic systems. The two maximum power point tracking controllers receive the partial derivative of power as inputs, and the output is the duty cycle corresponding to maximum power. A Photovoltaic generator with Boost converter is developed using MATLAB/Simulink to verify the preferences of the proposed techniques. SMC technique provides a good tracking speed in fast changing irradiation and when the irradiation changes slowly or is constant the panel power of FLC technique presents a much smoother signal with less fluctuations.

Keywords: fuzzy logic controller, maximum power point, photovoltaic system, tracker, sliding mode controller

Procedia PDF Downloads 548
9208 Performance Analysis of New Types of Reference Targets Based on Spaceborne and Airborne SAR Data

Authors: Y. S. Zhou, C. R. Li, L. L. Tang, C. X. Gao, D. J. Wang, Y. Y. Guo

Abstract:

Triangular trihedral corner reflector (CR) has been widely used as point target for synthetic aperture radar (SAR) calibration and image quality assessment. The additional “tip” of the triangular plate does not contribute to the reflector’s theoretical RCS and if it interacts with a perfectly reflecting ground plane, it will yield an increase of RCS at the radar bore-sight and decrease the accuracy of SAR calibration and image quality assessment. Regarding this problem, two types of CRs were manufactured. One was the hexagonal trihedral CR. It is a self-illuminating CR with relatively small plate edge length, while large edge length usually introduces unexpected edge diffraction error. The other was the triangular trihedral CR with extended bottom plate which considers the effect of ‘tip’ into the total RCS. In order to assess the performance of the two types of new CRs, flight campaign over the National Calibration and Validation Site for High Resolution Remote Sensors was carried out. Six hexagonal trihedral CRs and two bottom-extended trihedral CRs, as well as several traditional triangular trihedral CRs, were deployed. KOMPSAT-5 X-band SAR image was acquired for the performance analysis of the hexagonal trihedral CRs. C-band airborne SAR images were acquired for the performance analysis of the bottom-extended trihedral CRs. The analysis results showed that the impulse response function of both the hexagonal trihedral CRs and bottom-extended trihedral CRs were much closer to the ideal sinc-function than the traditional triangular trihedral CRs. The flight campaign results validated the advantages of new types of CRs and they might be useful in the future SAR calibration mission.

Keywords: synthetic aperture radar, calibration, corner reflector, KOMPSAT-5

Procedia PDF Downloads 276
9207 Robotic Assisted vs Traditional Laparoscopic Partial Nephrectomy Peri-Operative Outcomes: A Comparative Single Surgeon Study

Authors: Gerard Bray, Derek Mao, Arya Bahadori, Sachinka Ranasinghe

Abstract:

The EAU currently recommends partial nephrectomy as the preferred management for localised cT1 renal tumours, irrespective of surgical approach. With the advent of robotic assisted partial nephrectomy, there is growing evidence that warm ischaemia time may be reduced compared to the traditional laparoscopic approach. There is still no clear differences between the two approaches with regards to other peri-operative and oncological outcomes. Current limitations in the field denote the lack of single surgeon series to compare the two approaches as other studies often include multiple operators of different experience levels. To the best of our knowledge, this study is the first single surgeon series comparing peri-operative outcomes of robotic assisted and laparoscopic PN. The current study aims to reduce intra-operator bias while maintaining an adequate sample size to assess the differences in outcomes between the two approaches. We retrospectively compared patient demographics, peri-operative outcomes, and renal function derangements of all partial nephrectomies undertaken by a single surgeon with experience in both laparoscopic and robotic surgery. Warm ischaemia time, length of stay, and acute renal function deterioration were all significantly reduced with robotic partial nephrectomy, compared to laparoscopic nephrectomy. This study highlights the benefits of robotic partial nephrectomy. Further prospective studies with larger sample sizes would be valuable additions to the current literature.

Keywords: partial nephrectomy, robotic assisted partial nephrectomy, warm ischaemia time, peri-operative outcomes

Procedia PDF Downloads 142