Search results for: smart tools
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5215

Search results for: smart tools

3025 Model Order Reduction Using Hybrid Genetic Algorithm and Simulated Annealing

Authors: Khaled Salah

Abstract:

Model order reduction has been one of the most challenging topics in the past years. In this paper, a hybrid solution of genetic algorithm (GA) and simulated annealing algorithm (SA) are used to approximate high-order transfer functions (TFs) to lower-order TFs. In this approach, hybrid algorithm is applied to model order reduction putting in consideration improving accuracy and preserving the properties of the original model which are two important issues for improving the performance of simulation and computation and maintaining the behavior of the original complex models being reduced. Compared to conventional mathematical methods that have been used to obtain a reduced order model of high order complex models, our proposed method provides better results in terms of reducing run-time. Thus, the proposed technique could be used in electronic design automation (EDA) tools.

Keywords: genetic algorithm, simulated annealing, model reduction, transfer function

Procedia PDF Downloads 143
3024 Defining a Holistic Approach for Model-Based System Engineering: Paradigm and Modeling Requirements

Authors: Hycham Aboutaleb, Bruno Monsuez

Abstract:

Current systems complexity has reached a degree that requires addressing conception and design issues while taking into account all the necessary aspects. Therefore, one of the main challenges is the way complex systems are specified and designed. The exponential growing effort, cost and time investment of complex systems in modeling phase emphasize the need for a paradigm, a framework and a environment to handle the system model complexity. For that, it is necessary to understand the expectations of the human user of the model and his limits. This paper presents a generic framework for designing complex systems, highlights the requirements a system model needs to fulfill to meet human user expectations, and defines the refined functional as well as non functional requirements modeling tools needs to meet to be useful in model-based system engineering.

Keywords: system modeling, modeling language, modeling requirements, framework

Procedia PDF Downloads 532
3023 Socio-Cultural Adaptation Approach to Enhance Intercultural Collaboration and Learning

Authors: Fadoua Ouamani, Narjès Bellamine Ben Saoud, Henda Hajjami Ben Ghézala

Abstract:

In the last few years and over the last decades, there was a growing interest in the development of Computer Supported Collaborative Learning (CSCL) environments. However, the existing systems ignore the variety of learners and their socio-cultural differences, especially in the case of distant and networked learning. In fact, within such collaborative learning environments, learners from different socio-cultural backgrounds may interact together. These learners evolve within various cultures and social contexts and acquire different socio-cultural values and behaviors. Thus, they should be assisted while communicating and collaborating especially in an intercultural group. Besides, the communication and collaboration tools provided to each learner must depend on and be adapted to her/his socio-cultural profile. The main goal of this paper is to present the proposed socio-cultural adaptation approach based on and guided by ontologies to adapt CSCL environments to the socio-cultural profiles of its users (learners or others).

Keywords: CSCL, socio-cultural profile, adaptation, ontology

Procedia PDF Downloads 361
3022 A Web and Cloud-Based Measurement System Analysis Tool for the Automotive Industry

Authors: C. A. Barros, Ana P. Barroso

Abstract:

Any industrial company needs to determine the amount of variation that exists within its measurement process and guarantee the reliability of their data, studying the performance of their measurement system, in terms of linearity, bias, repeatability and reproducibility and stability. This issue is critical for automotive industry suppliers, who are required to be certified by the 16949:2016 standard (replaces the ISO/TS 16949) of International Automotive Task Force, defining the requirements of a quality management system for companies in the automotive industry. Measurement System Analysis (MSA) is one of the mandatory tools. Frequently, the measurement system in companies is not connected to the equipment and do not incorporate the methods proposed by the Automotive Industry Action Group (AIAG). To address these constraints, an R&D project is in progress, whose objective is to develop a web and cloud-based MSA tool. This MSA tool incorporates Industry 4.0 concepts, such as, Internet of Things (IoT) protocols to assure the connection with the measuring equipment, cloud computing, artificial intelligence, statistical tools, and advanced mathematical algorithms. This paper presents the preliminary findings of the project. The web and cloud-based MSA tool is innovative because it implements all statistical tests proposed in the MSA-4 reference manual from AIAG as well as other emerging methods and techniques. As it is integrated with the measuring devices, it reduces the manual input of data and therefore the errors. The tool ensures traceability of all performed tests and can be used in quality laboratories and in the production lines. Besides, it monitors MSAs over time, allowing both the analysis of deviations from the variation of the measurements performed and the management of measurement equipment and calibrations. To develop the MSA tool a ten-step approach was implemented. Firstly, it was performed a benchmarking analysis of the current competitors and commercial solutions linked to MSA, concerning Industry 4.0 paradigm. Next, an analysis of the size of the target market for the MSA tool was done. Afterwards, data flow and traceability requirements were analysed in order to implement an IoT data network that interconnects with the equipment, preferably via wireless. The MSA web solution was designed under UI/UX principles and an API in python language was developed to perform the algorithms and the statistical analysis. Continuous validation of the tool by companies is being performed to assure real time management of the ‘big data’. The main results of this R&D project are: MSA Tool, web and cloud-based; Python API; New Algorithms to the market; and Style Guide of UI/UX of the tool. The MSA tool proposed adds value to the state of the art as it ensures an effective response to the new challenges of measurement systems, which are increasingly critical in production processes. Although the automotive industry has triggered the development of this innovative MSA tool, other industries would also benefit from it. Currently, companies from molds and plastics, chemical and food industry are already validating it.

Keywords: automotive Industry, industry 4.0, Internet of Things, IATF 16949:2016, measurement system analysis

Procedia PDF Downloads 214
3021 Aluminum Based Hexaferrite and Reduced Graphene Oxide a Suitable Microwave Absorber for Microwave Application

Authors: Sanghamitra Acharya, Suwarna Datar

Abstract:

Extensive use of digital and smart communication createsprolong expose of unwanted electromagnetic (EM) radiations. This harmful radiation creates not only malfunctioning of nearby electronic gadgets but also severely affects a human being. So, a suitable microwave absorbing material (MAM) becomes a necessary urge in the field of stealth and radar technology. Initially, Aluminum based hexa ferrite was prepared by sol-gel technique and for carbon derived composite was prepared by the simple one port chemical reduction method. Finally, composite films of Poly (Vinylidene) Fluoride (PVDF) are prepared by simple gel casting technique. Present work demands that aluminum-based hexaferrite phase conjugated with graphene in PVDF matrix becomes a suitable candidate both in commercially important X and Ku band. The structural and morphological nature was characterized by X-Ray diffraction (XRD), Field emission-scanning electron microscope (FESEM) and Raman spectra which conforms that 30-40 nm particles are well decorated over graphene sheet. Magnetic force microscopy (MFM) and conducting force microscopy (CFM) study further conforms the magnetic and conducting nature of composite. Finally, shielding effectiveness (SE) of the composite film was studied by using Vector network analyzer (VNA) both in X band and Ku band frequency range and found to be more than 30 dB and 40 dB, respectively. As prepared composite films are excellent microwave absorbers.

Keywords: carbon nanocomposite, microwave absorbing material, electromagnetic shielding, hexaferrite

Procedia PDF Downloads 178
3020 Technology Computer Aided Design Simulation of Space Charge Limited Conduction in Polycrystalline Thin Films

Authors: Kunj Parikh, S. Bhattacharya, V. Natarajan

Abstract:

TCAD numerical simulation is one of the most tried and tested powerful tools for designing devices in semiconductor foundries worldwide. It has also been used to explain conduction in organic thin films where the processing temperature is often enough to make homogeneous samples (often imperfect, but homogeneously imperfect). In this report, we have presented the results of TCAD simulation in multi-grain thin films. The work has addressed the inhomogeneity in one dimension, but can easily be extended to two and three dimensions. The effect of grain boundaries has mainly been approximated as barriers located at the junction between two adjacent grains. The effect of the value of grain boundary barrier, the bulk traps, and the measurement temperature have been investigated.

Keywords: polycrystalline thin films, space charge limited conduction, Technology Computer-Aided Design (TCAD) simulation, traps

Procedia PDF Downloads 214
3019 HD-WSComp: Hypergraph Decomposition for Web Services Composition Based on QoS

Authors: Samah Benmerbi, Kamal Amroun, Abdelkamel Tari

Abstract:

The increasing number of Web service (WS)providers throughout the globe, have produced numerous Web services providing the same or similar functionality. Therefore, there is a need of tools developing the best answer of queries by selecting and composing services with total transparency. This paper reviews various QoS based Web service selection mechanisms and architectures which facilitate qualitatively optimal selection, in other fact Web service composition is required when a request cannot be fulfilled by a single web service. In such cases, it is preferable to integrate existing web services to satisfy user’s request. We introduce an automatic Web service composition method based on hypergraph decomposition using hypertree decomposition method. The problem of selection and the composition of the web services is transformed into a resolution in a hypertree by exploring the relations of dependency between web services to get composite web service via employing an execution order of WS satisfying global request.

Keywords: web service, web service selection, web service composition, QoS, hypergraph decomposition, BE hypergraph decomposition, hypertree resolution

Procedia PDF Downloads 510
3018 Urban Citizenship in a Sensor Rich Society

Authors: Mike Dee

Abstract:

Urban public spaces are sutured with a range of surveillance and sensor technologies that claim to enable new forms of ‘data based citizen participation’, but also increase the tendency for ‘function-creep’, whereby vast amounts of data are gathered, stored and analysed in a broad application of urban surveillance. This kind of monitoring and capacity for surveillance connects with attempts by civic authorities to regulate, restrict, rebrand and reframe urban public spaces. A direct consequence of the increasingly security driven, policed, privatised and surveilled nature of public space is the exclusion or ‘unfavourable inclusion’ of those considered flawed and unwelcome in the ‘spectacular’ consumption spaces of many major urban centres. In the name of urban regeneration, programs of securitisation, ‘gentrification’ and ‘creative’ and ‘smart’ city initiatives refashion public space as sites of selective inclusion and exclusion. In this context of monitoring and control procedures, in particular, children and young people’s use of space in parks, neighbourhoods, shopping malls and streets is often viewed as a threat to the social order, requiring various forms of remedial action. This paper suggests that cities, places and spaces and those who seek to use them, can be resilient in working to maintain and extend democratic freedoms and processes enshrined in Marshall’s concept of citizenship, calling sensor and surveillance systems to account. Such accountability could better inform the implementation of public policy around the design, build and governance of public space and also understandings of urban citizenship in the sensor saturated urban environment.

Keywords: citizenship, public space, surveillance, young people

Procedia PDF Downloads 449
3017 Mass Customization of Chemical Protective Clothing

Authors: Eugenija Strazdiene, Violeta Bytautaite, Daivute Krisciuniene

Abstract:

The object of the investigation is the suit for chemical protection, which totally covers human body together with breathing apparatus, breathing mask and helmet (JSC Ansell Protective Solutions Lithuania). The end users of such clothing are the members of rescue team – firefighters. During the presentation, the results of 3D scanning with stationary Human Solutions scanner and portable Artec Eva scanner will be compared on the basis of the efficiency of scanning procedure and scanning accuracy. Also, the possibilities to exporting scanned bodies into specialized CAD systems for suit design development and material consumption calculation will be analyzed. The necessity to understand and to implement corresponding clothing material properties during 3D visualization of garment on CAD systems will be presented. During the presentation, the outcomes of the project ‘Smart and Safe Work Wear Clothing SWW’ will be discussed. The project is carried out under the Interreg Baltic Sea Region Program as 2014-2020 European territorial cooperation objective. Thematic priority is Capacity for Innovation. The main goal of the project is to improve competitiveness and to increase business possibilities for work wear enterprises in the Baltic Sea Region. The project focuses on mass customization of products for various end users. It engages textile and clothing manufacturing technology researchers, work wear producers, end users, as well as national textile and clothing branch organizations in Finland, Lithuania, Latvia, Estonia and Poland.

Keywords: CAD systems, mass customization, 3D scanning, safe work wear

Procedia PDF Downloads 202
3016 Effect of Cost Control and Cost Reduction Techniques in Organizational Performance

Authors: Babatunde Akeem Lawal

Abstract:

In any organization, the primary aim is to maximize profit, but the major challenges facing them is the increase in cost of operation because of this there is increase in cost of production that could lead to inevitable cost control and cost reduction scheme which make it difficult for most organizations to operate at the cost efficient frontier. The study aims to critically examine and evaluate the application of cost control and cost reduction in organization performance and also to review budget as an effective tool of cost control and cost reduction. A descriptive survey research was adopted. A total number of 40 respondent retrieved were used for the study. The analysis of data collected was undertaken by applying appropriate statistical tools. Regression analysis was used to test the hypothesis with the use of SPSS. Based on the findings; it was evident that cost control has a positive impact on organizational performance and also the style of management has a positive impact on organizational performance.

Keywords: organization, cost reduction, cost control, performance, budget, profit

Procedia PDF Downloads 603
3015 The Problem of the Use of Learning Analytics in Distance Higher Education: An Analytical Study of the Open and Distance University System in Mexico

Authors: Ismene Ithai Bras-Ruiz

Abstract:

Learning Analytics (LA) is employed by universities not only as a tool but as a specialized ground to enhance students and professors. However, not all the academic programs apply LA with the same goal and use the same tools. In fact, LA is formed by five main fields of study (academic analytics, action research, educational data mining, recommender systems, and personalized systems). These fields can help not just to inform academic authorities about the situation of the program, but also can detect risk students, professors with needs, or general problems. The highest level applies Artificial Intelligence techniques to support learning practices. LA has adopted different techniques: statistics, ethnography, data visualization, machine learning, natural language process, and data mining. Is expected that any academic program decided what field wants to utilize on the basis of his academic interest but also his capacities related to professors, administrators, systems, logistics, data analyst, and the academic goals. The Open and Distance University System (SUAYED in Spanish) of the University National Autonomous of Mexico (UNAM), has been working for forty years as an alternative to traditional programs; one of their main supports has been the employ of new information and communications technologies (ICT). Today, UNAM has one of the largest network higher education programs, twenty-six academic programs in different faculties. This situation means that every faculty works with heterogeneous populations and academic problems. In this sense, every program has developed its own Learning Analytic techniques to improve academic issues. In this context, an investigation was carried out to know the situation of the application of LA in all the academic programs in the different faculties. The premise of the study it was that not all the faculties have utilized advanced LA techniques and it is probable that they do not know what field of study is closer to their program goals. In consequence, not all the programs know about LA but, this does not mean they do not work with LA in a veiled or, less clear sense. It is very important to know the grade of knowledge about LA for two reasons: 1) This allows to appreciate the work of the administration to improve the quality of the teaching and, 2) if it is possible to improve others LA techniques. For this purpose, it was designed three instruments to determinate the experience and knowledge in LA. These were applied to ten faculty coordinators and his personnel; thirty members were consulted (academic secretary, systems manager, or data analyst, and coordinator of the program). The final report allowed to understand that almost all the programs work with basic statistics tools and techniques, this helps the administration only to know what is happening inside de academic program, but they are not ready to move up to the next level, this means applying Artificial Intelligence or Recommender Systems to reach a personalized learning system. This situation is not related to the knowledge of LA, but the clarity of the long-term goals.

Keywords: academic improvements, analytical techniques, learning analytics, personnel expertise

Procedia PDF Downloads 128
3014 Scalable Performance Testing: Facilitating The Assessment Of Application Performance Under Substantial Loads And Mitigating The Risk Of System Failures

Authors: Solanki Ravirajsinh

Abstract:

In the software testing life cycle, failing to conduct thorough performance testing can result in significant losses for an organization due to application crashes and improper behavior under high user loads in production. Simulating large volumes of requests, such as 5 million within 5-10 minutes, is challenging without a scalable performance testing framework. Leveraging cloud services to implement a performance testing framework makes it feasible to handle 5-10 million requests in just 5-10 minutes, helping organizations ensure their applications perform reliably under peak conditions. Implementing a scalable performance testing framework using cloud services and tools like JMeter, EC2 instances (Virtual machine), cloud logs (Monitor errors and logs), EFS (File storage system), and security groups offers several key benefits for organizations. Creating performance test framework using this approach helps optimize resource utilization, effective benchmarking, increased reliability, cost savings by resolving performance issues before the application is released. In performance testing, a master-slave framework facilitates distributed testing across multiple EC2 instances to emulate many concurrent users and efficiently handle high loads. The master node orchestrates the test execution by coordinating with multiple slave nodes to distribute the workload. Slave nodes execute the test scripts provided by the master node, with each node handling a portion of the overall user load and generating requests to the target application or service. By leveraging JMeter's master-slave framework in conjunction with cloud services like EC2 instances, EFS, CloudWatch logs, security groups, and command-line tools, organizations can achieve superior scalability and flexibility in their performance testing efforts. In this master-slave framework, JMeter must be installed on both the master and each slave EC2 instance. The master EC2 instance functions as the "brain," while the slave instances operate as the "body parts." The master directs each slave to execute a specified number of requests. Upon completion of the execution, the slave instances transmit their results back to the master. The master then consolidates these results into a comprehensive report detailing metrics such as the number of requests sent, encountered errors, network latency, response times, server capacity, throughput, and bandwidth. Leveraging cloud services, the framework benefits from automatic scaling based on the volume of requests. Notably, integrating cloud services allows organizations to handle more than 5-10 million requests within 5 minutes, depending on the server capacity of the hosted website or application.

Keywords: identify crashes of application under heavy load, JMeter with cloud Services, Scalable performance testing, JMeter master and slave using cloud Services

Procedia PDF Downloads 27
3013 The Masterplan for the Urban Regeneration of the Heritage District of Msheireb Downtown Doha, State of Qatar

Authors: Raffaello Furlan

Abstract:

In the 21st century, the sustainable urban development of GCC-cities is challenged by inhabitants’ over-dependency on private-use vehicles. In turn, this habit has generated problems of urban inefficiency, contributing to traffic congestion, pollution, urban sprawling, fragmentation of the urban fabric, and various environmental and social challenges. In the context of Doha, the capital city of the State of Qatar, the over-dependency on private-use vehicles is justified by the lack of alternative public modes of transportation that support the need to connect fragmented urban districts and provide an effective solution to urban sprawl. Therefore, the current construction of the Qatar Metro Rail is offering the potential for investigating and defining a strategy for the sustainable urban development and/or urban regeneration of transit villages (TODs) in Qatar. Namely, the aim of this research study is (i) to investigate the development of transit villages (TODs) in the cultural-heritage district of Msheireb, Downtown Doha, (ii) to explore how the introduction of the new public transport system of Doha Metro can be effectively utilized as means of urban regeneration of the cultural core of the city, (iii) to propose a masterplan for TOD suitable for the district, suiting and responding to regional cultural and societal values. The findings reveal that the strategies for the sustainable urban regeneration of Msheireb are based on (i) the integration of land-use and multimodal transportation systems, (ii) the implementation of the public realm, and (iii) conservation of culture and urban identity.

Keywords: sustainable urbanism, smart growth, TODs, cultural district, Msheireb Downtown Doha

Procedia PDF Downloads 245
3012 A Generalized Framework for Adaptive Machine Learning Deployments in Algorithmic Trading

Authors: Robert Caulk

Abstract:

A generalized framework for adaptive machine learning deployments in algorithmic trading is introduced, tested, and released as open-source code. The presented software aims to test the hypothesis that recent data contains enough information to form a probabilistically favorable short-term price prediction. Further, the framework contains various adaptive machine learning techniques that are geared toward generating profit during strong trends and minimizing losses during trend changes. Results demonstrate that this adaptive machine learning approach is capable of capturing trends and generating profit. The presentation also discusses the importance of defining the parameter space associated with the dynamic training data-set and using the parameter space to identify and remove outliers from prediction data points. Meanwhile, the generalized architecture enables common users to exploit the powerful machinery while focusing on high-level feature engineering and model testing. The presentation also highlights common strengths and weaknesses associated with the presented technique and presents a broad range of well-tested starting points for feature set construction, target setting, and statistical methods for enforcing risk management and maintaining probabilistically favorable entry and exit points. The presentation also describes the end-to-end data processing tools associated with FreqAI, including automatic data fetching, data aggregation, feature engineering, safe and robust data pre-processing, outlier detection, custom machine learning and statistical tools, data post-processing, and adaptive training backtest emulation, and deployment of adaptive training in live environments. Finally, the generalized user interface is also discussed in the presentation. Feature engineering is simplified so that users can seed their feature sets with common indicator libraries (e.g. TA-lib, pandas-ta). The user also feeds data expansion parameters to fill out a large feature set for the model, which can contain as many as 10,000+ features. The presentation describes the various object-oriented programming techniques employed to make FreqAI agnostic to third-party libraries and external data sources. In other words, the back-end is constructed in such a way that users can leverage a broad range of common regression libraries (Catboost, LightGBM, Sklearn, etc) as well as common Neural Network libraries (TensorFlow, PyTorch) without worrying about the logistical complexities associated with data handling and API interactions. The presentation finishes by drawing conclusions about the most important parameters associated with a live deployment of the adaptive learning framework and provides the road map for future development in FreqAI.

Keywords: machine learning, market trend detection, open-source, adaptive learning, parameter space exploration

Procedia PDF Downloads 89
3011 Social Media as an Interactive Learning Tool Applied to Faculty of Tourism and Hotels, Fayoum University

Authors: Islam Elsayed Hussein

Abstract:

The aim of this paper is to discover the impact of students’ attitude towards social media and the skills required to adopt social media as a university e-learning (2.0) platform. In addition, it measures the effect of social media adoption on interactive learning effectiveness. The population of this study was students at Faculty of tourism and Hotels, Fayoum University. A questionnaire was used as a research instrument to collect data from respondents, which had been selected randomly. Data had been analyzed using quantitative data analysis method. Findings showed that the students have a positive attitude towards adopting social networking in the learning process and they have also good skills for effective use of social networking tools. In addition, adopting social media is effectively affecting the interactive learning environment.

Keywords: attitude, skills, e-learning 2.0, interactive learning, Egypt

Procedia PDF Downloads 525
3010 Innovative Business Education Pedagogy: A Case Study of Action Learning at NITIE, Mumbai

Authors: Sudheer Dhume, T. Prasad

Abstract:

There are distinct signs of Business Education losing its sheen. It is more so in developing countries. One of the reasons is the value addition at the end of 2 year MBA program is not matching with the requirements of present times and expectations of the students. In this backdrop, Pedagogy Innovation has become prerequisite for making our MBA programs relevant and useful. This paper is the description and analysis of innovative Action Learning pedagogical approach adopted by a group of faculty members at NITIE Mumbai. It not only promotes multidisciplinary research but also enhances integration of the functional areas skillsets in the students. The paper discusses the theoretical bases of this pedagogy and evaluates the effectiveness of it vis-à-vis conventional pedagogical tools. The evaluation research using Bloom’s taxonomy framework showed that this blended method of Business Education is much superior as compared to conventional pedagogy.

Keywords: action learning, blooms taxonomy, business education, innovation, pedagogy

Procedia PDF Downloads 270
3009 A Petri Net Model to Obtain the Throughput of Unreliable Production Lines in the Buffer Allocation Problem

Authors: Joselito Medina-Marin, Alexandr Karelin, Ana Tarasenko, Juan Carlos Seck-Tuoh-Mora, Norberto Hernandez-Romero, Eva Selene Hernandez-Gress

Abstract:

A production line designer faces with several challenges in manufacturing system design. One of them is the assignment of buffer slots in between every machine of the production line in order to maximize the throughput of the whole line, which is known as the Buffer Allocation Problem (BAP). The BAP is a combinatorial problem that depends on the number of machines and the total number of slots to be distributed on the production line. In this paper, we are proposing a Petri Net (PN) Model to obtain the throughput in unreliable production lines, based on PN mathematical tools and the decomposition method. The results obtained by this methodology are similar to those presented in previous works, and the number of machines is not a hard restriction.

Keywords: buffer allocation problem, Petri Nets, throughput, production lines

Procedia PDF Downloads 308
3008 A Fast Multi-Scale Finite Element Method for Geophysical Resistivity Measurements

Authors: Mostafa Shahriari, Sergio Rojas, David Pardo, Angel Rodriguez- Rozas, Shaaban A. Bakr, Victor M. Calo, Ignacio Muga

Abstract:

Logging-While Drilling (LWD) is a technique to record down-hole logging measurements while drilling the well. Nowadays, LWD devices (e.g., nuclear, sonic, resistivity) are mostly used commercially for geo-steering applications. Modern borehole resistivity tools are able to measure all components of the magnetic field by incorporating tilted coils. The depth of investigation of LWD tools is limited compared to the thickness of the geological layers. Thus, it is a common practice to approximate the Earth’s subsurface with a sequence of 1D models. For a 1D model, we can reduce the dimensionality of the problem using a Hankel transform. We can solve the resulting system of ordinary differential equations (ODEs) either (a) analytically, which results in a so-called semi-analytic method after performing a numerical inverse Hankel transform, or (b) numerically. Semi-analytic methods are used by the industry due to their high performance. However, they have major limitations, namely: -The analytical solution of the aforementioned system of ODEs exists only for piecewise constant resistivity distributions. For arbitrary resistivity distributions, the solution of the system of ODEs is unknown by today’s knowledge. -In geo-steering, we need to solve inverse problems with respect to the inversion variables (e.g., the constant resistivity value of each layer and bed boundary positions) using a gradient-based inversion method. Thus, we need to compute the corresponding derivatives. However, the analytical derivatives of cross-bedded formation and the analytical derivatives with respect to the bed boundary positions have not been published to the best of our knowledge. The main contribution of this work is to overcome the aforementioned limitations of semi-analytic methods by solving each 1D model (associated with each Hankel mode) using an efficient multi-scale finite element method. The main idea is to divide our computations into two parts: (a) offline computations, which are independent of the tool positions and we precompute only once and use them for all logging positions, and (b) online computations, which depend upon the logging position. With the above method, (a) we can consider arbitrary resistivity distributions along the 1D model, and (b) we can easily and rapidly compute the derivatives with respect to any inversion variable at a negligible additional cost by using an adjoint state formulation. Although the proposed method is slower than semi-analytic methods, its computational efficiency is still high. In the presentation, we shall derive the mathematical variational formulation, describe the proposed multi-scale finite element method, and verify the accuracy and efficiency of our method by performing a wide range of numerical experiments and comparing the numerical solutions to semi-analytic ones when the latest are available.

Keywords: logging-While-Drilling, resistivity measurements, multi-scale finite elements, Hankel transform

Procedia PDF Downloads 386
3007 Compensation of Power Quality Disturbances Using DVR

Authors: R. Rezaeipour

Abstract:

One of the key aspects of power quality improvement in power system is the mitigation of voltage sags/swells and flicker. Custom power devices have been known as the best tools for voltage disturbances mitigation as well as reactive power compensation. Dynamic voltage restorer (DVR) which is the most efficient and effective modern custom power device can provide the most commercial solution to solve several problems of power quality in distribution networks. This paper deals with analysis and simulation technique of DVR based on instantaneous power theory which is a quick control to detect signals. The main purpose of this work is to remove three important disturbances including voltage sags/swells and flicker. Simulation of the proposed method was carried out on two sample systems by using MATLAB software environment and the results of simulation show that the proposed method is able to provide desirable power quality in the presence of wide range of disturbances.

Keywords: DVR, power quality, voltage sags, voltage swells, flicker

Procedia PDF Downloads 345
3006 The Development of Speaking Using Folk Tales Based on Performance Activities for Early-Childhood Students

Authors: Ms Yaowaluck Ruampol

Abstract:

The research on the development of using folk tales based on performance activities aimed to (1) study the development of speaking skill for early-childhood students, (2) evaluate the development of speaking skill before and after speaking activities. Ten students of Kindergarten level 2, who have enrolled in the subject of the research for speaking development of semester 2 in 2013, were purposively selected as the research cohort. The research tools were lesson plans for speaking activities and pre-posttest for speaking development that were approved for content validity and reliability (IOC=.66-1.00,0.967). The research found that the development of speaking skill of the research samples before using performance activities on folk tales in developing speaking skill was in the normal high level. Additionally, the results revealed that the preschoolers after applying speaking skill on performance activities also imaginatively created their speaking skill.

Keywords: speaking development, folk tales, performance activities, communication engineering

Procedia PDF Downloads 291
3005 Supply Chain Control and Inventory Management in Garment Industry

Authors: Nisa Nur Duman, Sümeyya Kiliç

Abstract:

In global competition conditions, survival of the plants by obtaining competitive advantage relies on the effective usage of existing sources. By this way, the plants can minimize their costs without losing their quality. They also take advantage took advantage on their competitors and enlarge customer portfolio by increasing profit margins. Changing structure of market and customer demands also change the structure of the competition between companies. Furthermore, competition is not only between the companies. By this manner, supply chain and supply chain management get importance by considering company performances. Companies that want to survive, search the ways of decreasing costs and the ways of meeting customer expectations. One of the important tools for reaching these goals is inventory managemet. The best inventory management system is meeting the demands by considering plant goals.

Keywords: Supply chain, inventory management, apparel sector, garment industry

Procedia PDF Downloads 370
3004 Pre-Eliminary Design Adjustable Workstation for Piston Assembly Line Considering Anthropometric for Indonesian People

Authors: T. Yuri M. Zagloel, Inaki M. Hakim, Syarafi A. M.

Abstract:

Manufacturing process has been considered as one of the most important activity in business process. It correlates with productivity and quality of the product so industries could fulfill customer’s demand. With the increasing demand from customer, industries must improve their manufacturing ability such as shorten lead time and reduce wastes on their process. Lean manufacturing has been considered as one of the tools to waste elimination in manufacturing or service industri. Workforce development is one practice in lean manufacturing that can reduce waste generated from operator such as waste of unnecessary motion. Anthropometric approach is proposed to determine the recommended measurement in operator’s work area. The method will get some dimensions from Indonesia people that related to piston workstation. The result from this research can be obtained new design for the workarea considering ergonomic aspect.

Keywords: adjustable, anthropometric, ergonomic, waste

Procedia PDF Downloads 400
3003 Identifying Physiological Markers That Are Sensitive to Cognitive Load in Preschoolers

Authors: Priyashri Kamlesh Sridhar, Suranga Nanayakkara

Abstract:

Current frameworks in assessment follow lesson delivery and rely heavily on test performance or teacher’s observations. This, however, neglects the underlying cognitive load during the learning process. Identifying the pivotal points when the load occurs helps design effective pedagogies and tools that respond to learners’ cognitive state. There has been limited research on quantifying cognitive load in preschoolers, real-time. In this study, we recorded electrodermal activity and heart rate variability (HRV) from 10 kindergarteners performing executive function tasks and Johnson Woodcock test of cognitive abilities. Preliminary findings suggest that there are indeed sensitive task-dependent markers in skin conductance (number of SCRs and average amplitude of SCRs) and HRV (mean heart rate and low frequency component) captured during the learning process.

Keywords: early childhood, learning, methodologies, pedagogies

Procedia PDF Downloads 320
3002 Brand Content Optimization: A Major Challenge for Sellers on Marketplaces

Authors: Richardson Ciguene, Bertrand Marron, Nicolas Habert

Abstract:

Today, more and more consumers are purchasing their products and services online. At the same time, the penetration rate of very small and medium-sized businesses on marketplaces continues to increase, which has the direct impact of intensifying competition between sellers. Thus, only the best-optimized deals are ranked well by algorithms and are visible to consumers. However, it is almost impossible to know all the Brand Content rules and criteria established by marketplaces, which is essential to optimizing their product sheets, especially since these rules change constantly. In this paper, we propose to detail this question of Brand Content optimization by taking into account the case of Amazon in order to capture the scientific dimension behind such a subject. In a second step, we will present the genesis of our research project, DEEPERFECT, which aims to set up original methods and effective tools in order to help sellers present on marketplaces in the optimization of their branded content.

Keywords: e-commerce, scoring, marketplace, Amazon, brand content, product sheets

Procedia PDF Downloads 123
3001 Inflating the Public: A Series of Urban Interventions

Authors: Veronika Antoniou, Rene Carraz, Yiorgos Hadjichristou

Abstract:

The Green Urban Lab took the form of public installations that were placed at various locations in four cities in Cyprus. These installations - through which a series of events, activities, workshops and research took place - were the main tools in regenerating a series of urban public spaces in Cyprus. The purpose of this project was to identify issues and opportunities related to public space and to offer guidelines on how design and participatory democracy improvements could strengthen civil society, while raising the quality of the urban public scene. Giant inflatable structures were injected in important urban fragments in order to accommodate series of events. The design and playful installation generated a wide community engagement. The fluid presence of the installations acted as a catalyst for social interaction. They were accessed and viewed effortlessly and surprisingly, creating opportunities to rediscover public spaces.

Keywords: bottom-up initiatives, creativity, public space, social innovation, urban environments

Procedia PDF Downloads 516
3000 Effects of IPPC Permits on Ambient Air Quality

Authors: C. Cafaro, P. Ceci, L. De Giorgi

Abstract:

The aim of this paper is to give an assessment of environmental effects of IPPC permit conditions of installations that are in the specific territory with a high concentration of industrial activities. The IPPC permit is the permit that each operator should hold to operate the installation as stated by the directive 2010/75/UE on industrial emissions (integrated pollution prevention and control), known as IED (Industrial Emissions Directive). The IPPC permit includes all the measures necessary to achieve a high level of protection of the environment as a whole, also defining the monitoring requirements as measurement methodology, frequency, and evaluation procedure. The emissions monitoring of a specific plant may also give indications of the contribution of these emissions on the air quality of a definite area. So, it is clear that the IPPC permits are important tools both to improve the environmental framework and to achieve the air quality standards, assisting in assessing the possible industrial sources contributions to air pollution.

Keywords: IPPC, IED, emissions, permits, air quality, large combustion plants

Procedia PDF Downloads 450
2999 Application of Lean Manufacturing in Brake Shoe Manufacturing Plant: A Case Study

Authors: Anees K. Ahamed, Aakash Kumar R. G., Raj M. Mohan

Abstract:

The main objective is to apply lean tools to identify and eliminate waste in and among the work stations so as to improve the process speed and quality. From the top seven wastes in the lean concept, we consider the movement of materials, defects, and inventory for the improvement since these cause the major impact on the performance measures. The layout was improved to reduce the movement of materials. It also quantifies the reduction in movement among the work stations. Value stream mapping has been used for identification of waste. Cause and effect diagram and 5W analysis are used to identify the reasons for defects and to provide the counter measures. Some cycle time reduction techniques also proposed to improve the productivity. Lean Audit check sheet was also used to identify the current position of the industry and to identify the gap to make the industry Lean.

Keywords: cause and effect diagram, cycle time reduction, defects, lean, waste reduction

Procedia PDF Downloads 385
2998 Timing and Noise Data Mining Algorithm and Software Tool in Very Large Scale Integration (VLSI) Design

Authors: Qing K. Zhu

Abstract:

Very Large Scale Integration (VLSI) design becomes very complex due to the continuous integration of millions of gates in one chip based on Moore’s law. Designers have encountered numerous report files during design iterations using timing and noise analysis tools. This paper presented our work using data mining techniques combined with HTML tables to extract and represent critical timing/noise data. When we apply this data-mining tool in real applications, the running speed is important. The software employs table look-up techniques in the programming for the reasonable running speed based on performance testing results. We added several advanced features for the application in one industry chip design.

Keywords: VLSI design, data mining, big data, HTML forms, web, VLSI, EDA, timing, noise

Procedia PDF Downloads 254
2997 Preparation of Magnetothermally Responsive Polymer Multilayer Films for Controlled Release Applications from Surfaces

Authors: Eda Cagli, Irem Erel Goktepe

Abstract:

Externally triggered and effective release of therapeutics from polymer nanoplatforms is one of the key issues in cancer treatment. In this study, we aim to prepare polymer multilayer films which are stable at physiological conditions (little or no drug release) but release drug molecules at acidic pH and via application of AC magnetic field. First, novel stimuli responsive diblock copolymers composed of pH- and temperature-responsive blocks were synthesized. Then, block copolymer micelles with pH-responsive core and temperature responsive coronae will be obtained via pH-induced self-assembly of these block copolymers in aqueous environment. A model anticancer drug, e.g. Doxorubicin will be loaded in the micellar cores. Second, superparamagnetic nanoparticles will be synthesized. Magnetic nanoparticles and drug loaded block copolymer micelles will be used as building blocks to construct the multilayers. To mimic the acidic nature of the tumor tissues, Doxorubicin release from the micellar cores will be induced at acidic conditions. Moreover, Doxorubicin release from the multilayers will be facilitated via magnetothermal trigger. Application of AC magnetic field will induce the heating of magnetic nanoparticles resulting in an increase in the temperature of the polymer platform. This increase in temperature is expected to trigger conformational changes on the temperature-responsive micelle coronae and facilitate the release of Doxorubicin from the surface. Such polymer platform may find use in biomedical applications.

Keywords: layer-by-layer films, magnetothermal trigger, smart polymers, stimuli responsive

Procedia PDF Downloads 364
2996 Using Machine Learning to Extract Patient Data from Non-standardized Sports Medicine Physician Notes

Authors: Thomas Q. Pan, Anika Basu, Chamith S. Rajapakse

Abstract:

Machine learning requires data that is categorized into features that models train on. This topic is important to the field of sports medicine due to the many tools it provides to physicians such as diagnosis support and risk assessment. Physician note that healthcare professionals take are usually unclean and not suitable for model training. The objective of this study was to develop and evaluate an advanced approach for extracting key features from sports medicine data without the need for extensive model training or data labeling. An LLM (Large Language Model) was given a narrative (Physician’s Notes) and prompted to extract four features (details about the patient). The narrative was found in a datasheet that contained six columns: Case Number, Validation Age, Validation Gender, Validation Diagnosis, Validation Body Part, and Narrative. The validation columns represent the accurate responses that the LLM attempts to output. With the given narrative, the LLM would output its response and extract the age, gender, diagnosis, and injured body part with each category taking up one line. The output would then be cleaned, matched, and added to new columns containing the extracted responses. Five ways of checking the accuracy were used: unclear count, substring comparison, LLM comparison, LLM re-check, and hand-evaluation. The unclear count essentially represented the extractions the LLM missed. This can be also understood as the recall score ([total - false negatives] over total). The rest of these correspond to the precision score ([total - false positives] over total). Substring comparison evaluated the validation (X) and extracted (Y) columns’ likeness by checking if X’s results were a substring of Y's findings and vice versa. LLM comparison directly asked an LLM if the X and Y’s results were similar. LLM Re-check prompted the LLM to see if the extracted results can be found in the narrative. Lastly, A selection of 1,000 random narratives was also selected and hand-evaluated to give an estimate of how well the LLM-based feature extraction model performed. With a selection of 10,000 narratives, the LLM-based approach had a recall score of roughly 98%. However, the precision scores of the substring comparison and LLM comparison models were around 72% and 76% respectively. The reason for these low figures is due to the minute differences between answers. For example, the ‘chest’ is a part of the ‘upper trunk’ however, these models cannot detect that. On the other hand, the LLM re-check and subset of hand-tested narratives showed a precision score of 96% and 95%. If this subset is used to extrapolate the possible outcome of the whole 10,000 narratives, the LLM-based approach would be strong in both precision and recall. These results indicated that an LLM-based feature extraction model could be a useful way for medical data in sports to be collected and analyzed by machine learning models. Wide use of this method could potentially increase the availability of data thus improving machine learning algorithms and supporting doctors with more enhanced tools.

Keywords: AI, LLM, ML, sports

Procedia PDF Downloads 0