Search results for: user needs
370 The Impact of the Fitness Center Ownership Structure on the Service Quality Perception in the Fitness in Serbia
Authors: Dragan Zivotic, Mirjana Ilic, Aleksandra Perovic, Predrag Gavrilovic
Abstract:
As with the provision of other services, the service quality perception is one of the key factors that the modern manager must pay attention to. Countries in which the state regulation is in transition also have specific features in providing fitness services. Identification of the dimensions in which the most significant different service quality perception between different types of fitness centers, enables managers to profile the offer according to the wishes and expectations of users. The aim of the paper was the comparison of the quality of services perception in the field of fitness in Serbia between three categories of fitness centers: the privately owned centers, the publicly owned centers, and the Public-private partnership centers. For this research 350 respondents of both genders (174 men and 176 women) were interviewed, aged between 18 and 68 years, being beneficiaries of fitness services for at least 1 year. Administered questionnaire with 100 items provided information about the 15 basic areas in which they expressed the service quality perception in the gym. The core sample was composed of 212 service users in private fitness centers, 69 service users in public fitness centers and 69 service users in the public-private partnership. Sub-samples were equal in representation of women and men, as well as by age and length of use of fitness services. The obtained results were subject of univariate analysis with the Kruskal-Wallis non-parametric analysis of variance. Significant differences between the analyzed sub-samples were not found solely in the areas of rapid response and quality outcomes. In the multivariate model, the results were processed by backward stepwise discriminant analysis that extracted 3 areas that maximize the differences between sub-samples: material and technical basis, secondary facilities and coaches. By applying the classification function 93.87% of private centers services users, 62.32% of public centers services users and 85.51% of the public-private partnership centers users of services were correctly classified (total 86.00%). These results allow optimizing the allocation of the necessary resources in profiling offers of a fitness center in order to optimally adjust it to the user’s needs and expectations.Keywords: fitness, quality perception, management, public ownership, private ownership, public-private partnership, discriminative analysis
Procedia PDF Downloads 293369 Dynamic Web-Based 2D Medical Image Visualization and Processing Software
Authors: Abdelhalim. N. Mohammed, Mohammed. Y. Esmail
Abstract:
In the course of recent decades, medical imaging has been dominated by the use of costly film media for review and archival of medical investigation, however due to developments in networks technologies and common acceptance of a standard digital imaging and communication in medicine (DICOM) another approach in light of World Wide Web was produced. Web technologies successfully used in telemedicine applications, the combination of web technologies together with DICOM used to design a web-based and open source DICOM viewer. The Web server allowance to inquiry and recovery of images and the images viewed/manipulated inside a Web browser without need for any preinstalling software. The dynamic site page for medical images visualization and processing created by using JavaScript and HTML5 advancements. The XAMPP ‘apache server’ is used to create a local web server for testing and deployment of the dynamic site. The web-based viewer connected to multiples devices through local area network (LAN) to distribute the images inside healthcare facilities. The system offers a few focal points over ordinary picture archiving and communication systems (PACS): easy to introduce, maintain and independently platforms that allow images to display and manipulated efficiently, the system also user-friendly and easy to integrate with an existing system that have already been making use of web technologies. The wavelet-based image compression technique on which 2-D discrete wavelet transform used to decompose the image then wavelet coefficients are transmitted by entropy encoding after threshold to decrease transmission time, stockpiling cost and capacity. The performance of compression was estimated by using images quality metrics such as mean square error ‘MSE’, peak signal to noise ratio ‘PSNR’ and compression ratio ‘CR’ that achieved (83.86%) when ‘coif3’ wavelet filter is used.Keywords: DICOM, discrete wavelet transform, PACS, HIS, LAN
Procedia PDF Downloads 160368 Development of a Triangular Evaluation Protocol in a Multidisciplinary Design Process of an Ergometric Step
Authors: M. B. Ricardo De Oliveira, A. Borghi-Silva, E. Paravizo, F. Lizarelli, L. Di Thomazzo, D. Braatz
Abstract:
Prototypes are a critical feature in the product development process, as they help the project team visualize early concept flaws, communicate ideas and introduce an initial product testing. Involving stakeholders, such as consumers and users, in prototype tests allows the gathering of valuable feedback, contributing for a better product and making the design process more participatory. Even though recent studies have shown that user evaluation of prototypes is valuable, few articles provide a method or protocol on how designers should conduct it. This multidisciplinary study (involving the areas of physiotherapy, engineering and computer science) aims to develop an evaluation protocol, using an ergometric step prototype as the product prototype to be assessed. The protocol consisted of performing two tests (the 2 Minute Step Test and the Portability Test) to allow users (patients) and consumers (physiotherapists) to have an experience with the prototype. Furthermore, the protocol contained four Likert-Scale questionnaires (one for users and three for consumers), that inquired participants about how they perceived the design characteristics of the product (performance, safety, materials, maintenance, portability, usability and ergonomics), in their use of the prototype. Additionally, the protocol indicated the need to conduct interviews with the product designers, in order to link their feedback to the ones from the consumers and users. Both tests and interviews were recorded for further analysis. The participation criteria for the study was gender and age for patients, gender and experience with 2 Minute Step Test for physiotherapists and involvement level in the product development project for designers. The questionnaire's reliability was validated using Cronbach's Alpha and the quantitative data of the questionnaires were analyzed using non-parametric hypothesis tests with a significance level of 0.05 (p <0.05) and descriptive statistics. As a result, this study provides a concise evaluation protocol which can assist designers in their development process, collecting quantitative feedback from consumer and users, and qualitative feedback from designers.Keywords: Product Design, Product Evaluation, Prototypes, Step
Procedia PDF Downloads 117367 [Keynote Speech]: Curiosity, Innovation and Technological Advancements Shaping the Future of Science, Technology, Engineering and Mathematics Education
Authors: Ana Hol
Abstract:
We live in a constantly changing environment where technology has become an integral component of our day to day life. We rely heavily on mobile devices, we search for data via web, we utilise smart home sensors to create the most suited ambiences and we utilise applications to shop, research, communicate and share data. Heavy reliance on technology therefore is creating new connections between STEM (Science, Technology, Engineering and Mathematics) fields which in turn rises a question of what the STEM education of the future should be like? This study was based on the reviews of the six Australian Information Systems students who undertook an international study tour to India where they were given an opportunity to network, communicate and meet local students, staff and business representatives and from them learn about the local business implementations, local customs and regulations. Research identifies that if we are to continue to implement and utilise electronic devices on the global scale, such as for example implement smart cars that can smoothly cross borders, we will need the workforce that will have the knowledge about the cars themselves, their parts, roads and transport networks, road rules, road sensors, road monitoring technologies, graphical user interfaces, movement detection systems as well as day to day operations, legal rules and regulations of each region and country, insurance policies, policing and processes so that the wide array of sensors can be controlled across country’s borders. In conclusion, it can be noted that allowing students to learn about the local conditions, roads, operations, business processes, customs and values in different countries is giving students a cutting edge advantage as such knowledge cannot be transferred via electronic sources alone. However once understanding of each problem or project is established, multidisciplinary innovative STEM projects can be smoothly conducted.Keywords: STEM, curiosity, innovation, advancements
Procedia PDF Downloads 199366 Post Mutiny Institutional Buildings of India: A Visual Language of Reconciliation
Authors: Aruna Ramani Grover
Abstract:
In 1857 the British army in Bengal rose in mutiny. The outcome of a yearlong stifle was the abolition of the East India Company and establishment of Crown rule in 1958. Despite being a political democracy back home, with the declaration of Queen Victoria as Empress of India, the British established themselves as imperialistic successors to the Mughals in India. With the institution of the Crown role in the sub-continent, there was a serious endeavour for systematic governance. This led to infrastructure development and creation of institutions as well. The outcome was many public buildings like courts, railway- stations and headquarters, senates, post offices, banks, libraries, memorial halls, museums, memorials, theatres, government colleges, residential-schools and clock towers. These were built in the old and emerging urban settlements of the sub-continents. In the realm of architecture, like all political masters of the past, the British architects too encountered the living tradition of the sub-continent. A bewildering plethora of buildings in various climatic zones, using local materials and crafted with tools and techniques of the region by local craftsmen had to be understood and assimilated. The19th century British architects who built in India, designed institutional buildings which were functional and responded to the need of the user. In visual terms however, it was a completely different story. Using the manifest elements of the complex and layered indigenous tradition, they fashioned buildings to create an architecture of reconciliation. Treating the traditional architecture as a pattern book, finished buildings was served up to the local population coloured by their understanding of tradition. This paper will discuss a series of building some of which are the Senate House and Law courts at Madras, the Municipal Building and Gateway of India in Bombay, the Muir college in Allahabad, Mayo college Ajmer, the Mubarak Mahal in Jaipur to demonstrate how a visual language of reconciliation was created.Keywords: infrastructure, British architects, tradition, pattern book, reconciliation
Procedia PDF Downloads 32365 Montelukast Doesn’t Decrease the Risk of Cardiovascular Disease in Asthma Patients in Taiwan
Authors: Sheng Yu Chen, Shi-Heng Wang
Abstract:
Aim: Based on human, animal experiments, and genetic studies, cysteinyl leukotrienes, LTC4, LTD4, and LTE4, are inflammatory substances that are metabolized by 5-lipooxygenase from arachidonic acid, and these substances trigger asthma. In addition, the synthetic pathway of cysteinyl leukotriene is relevant to the increase in cardiovascular diseases such as myocardial ischemia and stroke. Given the situation, we aim to investigate whether cysteinyl leukotrienes receptor antagonist (LTRA), montelukast which cures those who have asthma has potential protective effects on cardiovascular diseases. Method: We conducted a cohort study, and enrolled participants which are newly diagnosed with asthma (ICD-9 CM code 493. X) between 2002 to 2011. The data source is from Taiwan National Health Insurance Research Database Patients with a previous history of myocardial infarction or ischemic stroke were excluded. Among the remaining participants, every montelukast user was matched with two randomly non-users by sex, and age. The incident cardiovascular diseases, including myocardial infarction and ischemic stroke, were regarded as outcomes. We followed the participants until outcomes come first or the end of the following period. To explore the protective effect of montelukast on the risk of cardiovascular disease, we use multivariable Cox regression to estimate the hazard ratio with adjustment for potential confounding factors. Result: There are 55876 newly diagnosed asthma patients who had at least one claim of inpatient admission or at least three claims of outpatient records. We enrolled 5350 montelukast users and 10700 non-users in this cohort study. The following mean (±SD) time of the Montelukast group is 5 (±2.19 )years, and the non-users group is 6.2 5.47 (± 2.641) years. By using multivariable Cox regression, our analysis indicated that the risk of incident cardiovascular diseases between montelukast users (n=43, 0.8%) and non-users (n=111, 1.04%) is approximately equal. [adjusted hazard ratio 0.992; P-value:0.9643] Conclusion: In this population-based study, we found that the use of montelukast is not associated with a decrease in incident MI or IS.Keywords: asthma, inflammation, montelukast, insurance research database, cardiovascular diseases
Procedia PDF Downloads 82364 Wireless FPGA-Based Motion Controller Design by Implementing 3-Axis Linear Trajectory
Authors: Kiana Zeighami, Morteza Ozlati Moghadam
Abstract:
Designing a high accuracy and high precision motion controller is one of the important issues in today’s industry. There are effective solutions available in the industry but the real-time performance, smoothness and accuracy of the movement can be further improved. This paper discusses a complete solution to carry out the movement of three stepper motors in three dimensions. The objective is to provide a method to design a fully integrated System-on-Chip (SOC)-based motion controller to reduce the cost and complexity of production by incorporating Field Programmable Gate Array (FPGA) into the design. In the proposed method the FPGA receives its commands from a host computer via wireless internet communication and calculates the motion trajectory for three axes. A profile generator module is designed to realize the interpolation algorithm by translating the position data to the real-time pulses. This paper discusses an approach to implement the linear interpolation algorithm, since it is one of the fundamentals of robots’ movements and it is highly applicable in motion control industries. Along with full profile trajectory, the triangular drive is implemented to eliminate the existence of error at small distances. To integrate the parallelism and real-time performance of FPGA with the power of Central Processing Unit (CPU) in executing complex and sequential algorithms, the NIOS II soft-core processor was added into the design. This paper presents different operating modes such as absolute, relative positioning, reset and velocity modes to fulfill the user requirements. The proposed approach was evaluated by designing a custom-made FPGA board along with a mechanical structure. As a result, a precise and smooth movement of stepper motors was observed which proved the effectiveness of this approach.Keywords: 3-axis linear interpolation, FPGA, motion controller, micro-stepping
Procedia PDF Downloads 208363 Heterogeneous-Resolution and Multi-Source Terrain Builder for CesiumJS WebGL Virtual Globe
Authors: Umberto Di Staso, Marco Soave, Alessio Giori, Federico Prandi, Raffaele De Amicis
Abstract:
The increasing availability of information about earth surface elevation (Digital Elevation Models DEM) generated from different sources (remote sensing, Aerial Images, Lidar) poses the question about how to integrate and make available to the most than possible audience this huge amount of data. In order to exploit the potential of 3D elevation representation the quality of data management plays a fundamental role. Due to the high acquisition costs and the huge amount of generated data, highresolution terrain surveys tend to be small or medium sized and available on limited portion of earth. Here comes the need to merge large-scale height maps that typically are made available for free at worldwide level, with very specific high resolute datasets. One the other hand, the third dimension increases the user experience and the data representation quality, unlocking new possibilities in data analysis for civil protection, real estate, urban planning, environment monitoring, etc. The open-source 3D virtual globes, which are trending topics in Geovisual Analytics, aim at improving the visualization of geographical data provided by standard web services or with proprietary formats. Typically, 3D Virtual globes like do not offer an open-source tool that allows the generation of a terrain elevation data structure starting from heterogeneous-resolution terrain datasets. This paper describes a technological solution aimed to set up a so-called “Terrain Builder”. This tool is able to merge heterogeneous-resolution datasets, and to provide a multi-resolution worldwide terrain services fully compatible with CesiumJS and therefore accessible via web using traditional browser without any additional plug-in.Keywords: Terrain Builder, WebGL, Virtual Globe, CesiumJS, Tiled Map Service, TMS, Height-Map, Regular Grid, Geovisual Analytics, DTM
Procedia PDF Downloads 425362 Exploring the Perspective of Service Quality in mHealth Services during the COVID-19 Pandemic
Authors: Wan-I Lee, Nelio Mendoza Figueredo
Abstract:
The impact of COVID-19 has a significant effect on all sectors of society globally. Health information technology (HIT) has become an effective health strategy in this age of distancing. In this regard, Mobile Health (mHealth) plays a critical role in managing patient and provider workflows during the COVID-19 pandemic. Therefore, the users' perception of service quality about mHealth services plays a significant role in shaping confidence and subsequent behaviors regarding the mHealth users' intention of use. This study's objective was to explore levels of user attributes analyzed by a qualitative method of how health practitioners and patients are satisfied or dissatisfied with using mHealth services; and analyzed the users' intention in the context of Taiwan during the COVID-19 pandemic. This research explores the experienced usability of a mHealth services during the Covid-19 pandemic. This study uses qualitative methods that include in-depth and semi-structured interviews that investigate participants' perceptions and experiences and the meanings they attribute to them. The five cases consisted of health practitioners, clinic staff, and patients' experiences using mHealth services. This study encourages participants to discuss issues related to the research question by asking open-ended questions, usually in one-to-one interviews. The findings show the positive and negative attributes of mHealth service quality. Hence, the significant importance of patients' and health practitioners' issues on several dimensions of perceived service quality is system quality, information quality, and interaction quality. A concept map for perceptions regards to emergency uses' intention of mHealth services process is depicted. The findings revealed that users pay more attention to "Medical care", "ease of use" and "utilitarian benefits" and have less importance for "Admissions and Convenience" and "Social influence". To improve mHealth services, the mHealth providers and health practitioners should better manage users' experiences to enhance mHealth services. This research contributes to the understanding of service quality issues in mHealth services during the COVID-19 pandemic.Keywords: COVID-19, mobile health, service quality, use intention
Procedia PDF Downloads 147361 Utilizing Dowel-Laminated Mass Timber Components in Residential Multifamily Structures: A Case Study
Authors: Theodore Panton
Abstract:
As cities in the United States experience critical housing shortages, mass timber presents the opportunity to address this crisis in housing supply while taking advantage of the carbon-positive benefits of sustainably forested wood fiber. Mass timber, however, currently has a low level of adoption in residential multifamily structures due to the risk-averse nature of change within the construction financing, Architecture / Engineering / Contracting (AEC) communities, as well as various agency approval challenges. This study demonstrates how mass timber can be used within the cost and feasibility parameters of a typical multistory residential structure and ultimately address the need for dense urban housing. This study will utilize The Garden District, a mixed-use market-rate housing project in Woodinville, Washington, as a case study to illuminate the potential of mass timber in this application. The Garden District is currently in final stages of permit approval and will commence construction in 2023. It will be the tallest dowel-laminated timber (DLT) residential structure in the United States when completed. This case study includes economic, technical, and design reference points to demonstrate the relevance of the use of this system and its ability to deliver “triple bottom line” results. In terms of results, the study establishes scalable and repeatable approaches to project design and delivery of mass timber in multifamily residential uses and includes economic data, technical solutions, and a summary of end-user advantages. This study discusses the third party tested systems for satisfying acoustical requirements within dwelling units, a key to resolving the use of mass timber within multistory residential use. Lastly, the study will also compare the mass timber solution with a comparable cold formed steel (CFS) system with a similar program, which indicates a net carbon savings of over three million tons over the life cycle of the building.Keywords: DLT, dowell laminated timber, mass timber, market rate multifamily
Procedia PDF Downloads 121360 Reliability Analysis of Glass Epoxy Composite Plate under Low Velocity
Authors: Shivdayal Patel, Suhail Ahmad
Abstract:
Safety assurance and failure prediction of composite material component of an offshore structure due to low velocity impact is essential for associated risk assessment. It is important to incorporate uncertainties associated with material properties and load due to an impact. Likelihood of this hazard causing a chain of failure events plays an important role in risk assessment. The material properties of composites mostly exhibit a scatter due to their in-homogeneity and anisotropic characteristics, brittleness of the matrix and fiber and manufacturing defects. In fact, the probability of occurrence of such a scenario is due to large uncertainties arising in the system. Probabilistic finite element analysis of composite plates due to low-velocity impact is carried out considering uncertainties of material properties and initial impact velocity. Impact-induced damage of composite plate is a probabilistic phenomenon due to a wide range of uncertainties arising in material and loading behavior. A typical failure crack initiates and propagates further into the interface causing de-lamination between dissimilar plies. Since individual crack in the ply is difficult to track. The progressive damage model is implemented in the FE code by a user-defined material subroutine (VUMAT) to overcome these problems. The limit state function is accordingly established while the stresses in the lamina are such that the limit state function (g(x)>0). The Gaussian process response surface method is presently adopted to determine the probability of failure. A comparative study is also carried out for different combination of impactor masses and velocities. The sensitivity based probabilistic design optimization procedure is investigated to achieve better strength and lighter weight of composite structures. Chain of failure events due to different modes of failure is considered to estimate the consequences of failure scenario. Frequencies of occurrence of specific impact hazards yield the expected risk due to economic loss.Keywords: composites, damage propagation, low velocity impact, probability of failure, uncertainty modeling
Procedia PDF Downloads 279359 The Sociology of the Facebook: An Exploratory Study
Authors: Liana Melissa E. de la Rosa, Jayson P. Ada
Abstract:
This exploratory study was conducted to determine the sociology of the Facebook. Specifically, it aimed to know the socio-demographic profile of the respondents in terms of age, sex, year level and monthly allowance; find out the common usage of Facebook to the respondents; identify the features of Facebook that are commonly used by the respondents; understand the benefits and risks of using the Facebook; determine how frequent the respondents use the Facebook; and find out if there is a significant relationship between socio-demographic profile of the respondents and their Facebook usage. This study used the exploratory research design and correlational design employing research survey questionnaire as its main data gathering instrument. Students of the University of Eastern Philippines were selected as the respondents of this study through quota sampling. Ten (10) students were randomly selected from each college of the university. Based on the findings of this study, the following conclusion were drawn: The majority of the respondents are aged 18 and 21 old, female, are third year students, and have monthly allowance of P 2,000 above. On the respondents’ usage of Facebook, the majority of use the Facebook on a daily basis for one to two (1-2) hours everyday. And most users used Facebook by renting a computer in an internet cafe. On the use of Facebook, most users have created their profiles mainly to connect with people and gain new friends. The most commonly used features of Facebook, are: photos application, like button, wall, notification, friend, chat, network, groups and “like” pages status updates, messages and inbox and events. While the other Facebook features that are seldom used by the respondents are games, news feed, user name, video sharing and notes. And the least used Facebook features are questions, poke feature, credits and the market place. The respondents stated that the major benefit that the Facebook has given to its users is its ability to keep in touch with family members or friends while the main risk identified is that the users can become addicted to the Internet. On the tests of relationships between the respondents’ use of Facebook and the four (4) socio-demographic profile variables: age, sex, year level, and month allowance, were found to be not significantly related to the respondents’ use of the Facebook. While the variable found to be significantly related was gender.Keywords: Facebook, sociology, social networking, exploratory study
Procedia PDF Downloads 289358 Health Economics in the Cost-Benefit Analysis of Transport Schemes
Authors: Henry Kelly, Helena Shaw
Abstract:
This paper will seek how innovative methods from Health Economics and, to a lesser extent, wellbeing analysis can be applied in the Cost-Benefit Analysis (CBA) of transport infrastructure and policy interventions. The context for this will focus on the framework articulated by the UK Treasury (finance department) and the English Department for Transport. Both have well-established methods for undertaking CBA, but there is increased policy interest, particularly at a regional level of exploring broader strategic goals beyond those traditionally associated with transport user benefits, productivity gains, and labour market access. Links to different CBA approaches internationally, such as New Zealand, France, and Wales will be referenced. By exploring a complementary method of accessing the impacts of policies through the quantification of health impacts is a fruitful line to explore. In a previous piece of work, 14 impact pathways were identified, mapping the relationship between transport and health. These are wide-ranging, from improved employment prospects, the stress of unreliable journey times, and air quality to isolation and loneliness. Importantly, we will consider these different measures of health from an intersectional point of view to ensure that the basis that remains in the health industry does not get translated across to this work. The objective is to explore how a CBA based on these pathways may, through quantifying forecast impacts in terms of Quality-Adjusted Life Years may, produce different findings than a standard approach. Of particular interest is how a health-based approach may have different distributional impacts on socio-economic groups and may favour distinct types of interventions. Consideration will be given to the degree this approach may double-count impacts or if it is possible to identify additional benefits to the established CBA approach. The investigation will explore a range of schemes, from a high-speed rail link, highway improvements, rural mobility hubs, and coach services to cycle lanes. The conclusions should aid the progression of methods concerning the assessment of publicly funded infrastructure projects.Keywords: cost-benefit analysis, health, QALYs transport
Procedia PDF Downloads 80357 Computation of Radiotherapy Treatment Plans Based on CT to ED Conversion Curves
Authors: B. Petrović, L. Rutonjski, M. Baucal, M. Teodorović, O. Čudić, B. Basarić
Abstract:
Radiotherapy treatment planning computers use CT data of the patient. For the computation of a treatment plan, treatment planning system must have an information on electron densities of tissues scanned by CT. This information is given by the conversion curve CT (CT number) to ED (electron density), or simply calibration curve. Every treatment planning system (TPS) has built in default CT to ED conversion curves, for the CTs of different manufacturers. However, it is always recommended to verify the CT to ED conversion curve before actual clinical use. Objective of this study was to check how the default curve already provided matches the curve actually measured on a specific CT, and how much it influences the calculation of a treatment planning computer. The examined CT scanners were from the same manufacturer, but four different scanners from three generations. The measurements of all calibration curves were done with the dedicated phantom CIRS 062M Electron Density Phantom. The phantom was scanned, and according to real HU values read at the CT console computer, CT to ED conversion curves were generated for different materials, for same tube voltage 140 kV. Another phantom, CIRS Thorax 002 LFC which represents an average human torso in proportion, density and two-dimensional structure, was used for verification. The treatment planning was done on CT slices of scanned CIRS LFC 002 phantom, for selected cases. Interest points were set in the lungs, and in the spinal cord, and doses recorded in TPS. The overall calculated treatment times for four scanners and default scanner did not differ more than 0.8%. Overall interest point dose in bone differed max 0.6% while for single fields was maximum 2.7% (lateral field). Overall interest point dose in lungs differed max 1.1% while for single fields was maximum 2.6% (lateral field). It is known that user should verify the CT to ED conversion curve, but often, developing countries are facing lack of QA equipment, and often use default data provided. We have concluded that the CT to ED curves obtained differ in certain points of a curve, generally in the region of higher densities. This influences the treatment planning result which is not significant, but definitely does make difference in the calculated dose.Keywords: Computation of treatment plan, conversion curve, radiotherapy, electron density
Procedia PDF Downloads 486356 Testing and Validation Stochastic Models in Epidemiology
Authors: Snigdha Sahai, Devaki Chikkavenkatappa Yellappa
Abstract:
This study outlines approaches for testing and validating stochastic models used in epidemiology, focusing on the integration and functional testing of simulation code. It details methods for combining simple functions into comprehensive simulations, distinguishing between deterministic and stochastic components, and applying tests to ensure robustness. Techniques include isolating stochastic elements, utilizing large sample sizes for validation, and handling special cases. Practical examples are provided using R code to demonstrate integration testing, handling of incorrect inputs, and special cases. The study emphasizes the importance of both functional and defensive programming to enhance code reliability and user-friendliness.Keywords: computational epidemiology, epidemiology, public health, infectious disease modeling, statistical analysis, health data analysis, disease transmission dynamics, predictive modeling in health, population health modeling, quantitative public health, random sampling simulations, randomized numerical analysis, simulation-based analysis, variance-based simulations, algorithmic disease simulation, computational public health strategies, epidemiological surveillance, disease pattern analysis, epidemic risk assessment, population-based health strategies, preventive healthcare models, infection dynamics in populations, contagion spread prediction models, survival analysis techniques, epidemiological data mining, host-pathogen interaction models, risk assessment algorithms for disease spread, decision-support systems in epidemiology, macro-level health impact simulations, socioeconomic determinants in disease spread, data-driven decision making in public health, quantitative impact assessment of health policies, biostatistical methods in population health, probability-driven health outcome predictions
Procedia PDF Downloads 6355 On Cloud Computing: A Review of the Features
Authors: Assem Abdel Hamed Mousa
Abstract:
The Internet of Things probably already influences your life. And if it doesn’t, it soon will, say computer scientists; Ubiquitous computing names the third wave in computing, just now beginning. First were mainframes, each shared by lots of people. Now we are in the personal computing era, person and machine staring uneasily at each other across the desktop. Next comes ubiquitous computing, or the age of calm technology, when technology recedes into the background of our lives. Alan Kay of Apple calls this "Third Paradigm" computing. Ubiquitous computing is essentially the term for human interaction with computers in virtually everything. Ubiquitous computing is roughly the opposite of virtual reality. Where virtual reality puts people inside a computer-generated world, ubiquitous computing forces the computer to live out here in the world with people. Virtual reality is primarily a horse power problem; ubiquitous computing is a very difficult integration of human factors, computer science, engineering, and social sciences. The approach: Activate the world. Provide hundreds of wireless computing devices per person per office, of all scales (from 1" displays to wall sized). This has required new work in operating systems, user interfaces, networks, wireless, displays, and many other areas. We call our work "ubiquitous computing". This is different from PDA's, dynabooks, or information at your fingertips. It is invisible; everywhere computing that does not live on a personal device of any sort, but is in the woodwork everywhere. The initial incarnation of ubiquitous computing was in the form of "tabs", "pads", and "boards" built at Xerox PARC, 1988-1994. Several papers describe this work, and there are web pages for the Tabs and for the Boards (which are a commercial product now): Ubiquitous computing will drastically reduce the cost of digital devices and tasks for the average consumer. With labor intensive components such as processors and hard drives stored in the remote data centers powering the cloud , and with pooled resources giving individual consumers the benefits of economies of scale, monthly fees similar to a cable bill for services that feed into a consumer’s phone.Keywords: internet, cloud computing, ubiquitous computing, big data
Procedia PDF Downloads 382354 3D Codes for Unsteady Interaction Problems of Continuous Mechanics in Euler Variables
Authors: M. Abuziarov
Abstract:
The designed complex is intended for the numerical simulation of fast dynamic processes of interaction of heterogeneous environments susceptible to the significant formability. The main challenges in solving such problems are associated with the construction of the numerical meshes. Currently, there are two basic approaches to solve this problem. One is using of Lagrangian or Lagrangian Eulerian grid associated with the boundaries of media and the second is associated with the fixed Eulerian mesh, boundary cells of which cut boundaries of the environment medium and requires the calculation of these cut volumes. Both approaches require the complex grid generators and significant time for preparing the code’s data for simulation. In this codes these problems are solved using two grids, regular fixed and mobile local Euler Lagrange - Eulerian (ALE approach) accompanying the contact and free boundaries, the surfaces of shock waves and phase transitions, and other possible features of solutions, with mutual interpolation of integrated parameters. For modeling of both liquids and gases, and deformable solids the Godunov scheme of increased accuracy is used in Lagrangian - Eulerian variables, the same for the Euler equations and for the Euler- Cauchy, describing the deformation of the solid. The increased accuracy of the scheme is achieved by using 3D spatial time dependent solution of the discontinuity problem (3D space time dependent Riemann's Problem solver). The same solution is used to calculate the interaction at the liquid-solid surface (Fluid Structure Interaction problem). The codes does not require complex 3D mesh generators, only the surfaces of the calculating objects as the STL files created by means of engineering graphics are given by the user, which greatly simplifies the preparing the task and makes it convenient to use directly by the designer at the design stage. The results of the test solutions and applications related to the generation and extension of the detonation and shock waves, loading the constructions are presented.Keywords: fluid structure interaction, Riemann's solver, Euler variables, 3D codes
Procedia PDF Downloads 439353 Slowness in Architecture: The Pace of Human Engagement with the Built Environment
Authors: Jaidev Tripathy
Abstract:
A human generation’s lifestyle, behaviors, habits, and actions are governed heavily by homogenous mindsets. But the current scenario is witnessing a rapid gap in this homogeneity as a result of an intervention, or rather, the dominance of the digital revolution in the human lifestyle. The current mindset for mass production, employment, multi-tasking, rapid involvement, and stiff competition to stay above the rest has led to a major shift in human consciousness. Architecture, as an entity, is being perceived differently. The screens are replacing the skies. The pace at which operation and evolution is taking place has increased. It is paradoxical, that time seems to be moving faster despite the intention to save time. Parallelly, there is an evident shift in architectural typologies spanning across different generations. The architecture of today is now seems influenced heavily from here and there. Mass production of buildings and over-exploitation of resources giving shape to uninspiring algorithmic designs, ambiguously catering to multiple user groups, has become a prevalent theme. Borrow-and-steal replaces influence, and the diminishing depth in today’s designs reflects a lack of understanding and connection. The digitally dominated world, perceived as an aid to connect and network, is making humans less capable of real-life interactions and understanding. It is not wrong, but it doesn’t seem right either. The engagement level between human beings and the built environment is a concern which surfaces. This leads to a question: Does human engagement drive architecture, or does architecture drive human engagement? This paper attempts to relook at architecture's capacity and its relativity with pace to influence the conscious decisions of a human being. Secondary research, supported with case examples, helps in understanding the translation of human engagement with the built environment through physicality of architecture. The procedure, or theme, is pace and the role of slowness in the context of human behaviors, thus bridging the widening gap between the human race and the architecture themselves give shape to, avoiding a possible future dystopian world.Keywords: junkspace, pace, perception, slowness
Procedia PDF Downloads 109352 System Dietadhoc® - A Fusion of Human-Centred Design and Agile Development for the Explainability of AI Techniques Based on Nutritional and Clinical Data
Authors: Michelangelo Sofo, Giuseppe Labianca
Abstract:
In recent years, the scientific community's interest in the exploratory analysis of biomedical data has increased exponentially. Considering the field of research of nutritional biologists, the curative process, based on the analysis of clinical data, is a very delicate operation due to the fact that there are multiple solutions for the management of pathologies in the food sector (for example can recall intolerances and allergies, management of cholesterol metabolism, diabetic pathologies, arterial hypertension, up to obesity and breathing and sleep problems). In this regard, in this research work a system was created capable of evaluating various dietary regimes for specific patient pathologies. The system is founded on a mathematical-numerical model and has been created tailored for the real working needs of an expert in human nutrition using the human-centered design (ISO 9241-210), therefore it is in step with continuous scientific progress in the field and evolves through the experience of managed clinical cases (machine learning process). DietAdhoc® is a decision support system nutrition specialists for patients of both sexes (from 18 years of age) developed with an agile methodology. Its task consists in drawing up the biomedical and clinical profile of the specific patient by applying two algorithmic optimization approaches on nutritional data and a symbolic solution, obtained by transforming the relational database underlying the system into a deductive database. For all three solution approaches, particular emphasis has been given to the explainability of the suggested clinical decisions through flexible and customizable user interfaces. Furthermore, the system has multiple software modules based on time series and visual analytics techniques that allow to evaluate the complete picture of the situation and the evolution of the diet assigned for specific pathologies.Keywords: medical decision support, physiological data extraction, data driven diagnosis, human centered AI, symbiotic AI paradigm
Procedia PDF Downloads 23351 The Development of the First Inter-Agency Residential Rehabilitation Service for Gambling Disorder with Complex Clinical Needs
Authors: Dragos Dragomir-Stanciu, Leon Marsh
Abstract:
Background As a response to the gaps identified in recent research in the provision of residential care to address co-occurring health needs, including mental health problems and complexities Gamble Aware has facilitated the possibility to provide a new service which would extend the NGTS provision of residential rehabilitation for gambling disorder with complex and co-morbid presentation. Gordon Moody, together with Adferiad have been successful in securing the tender for this service and this presentation aims to introduce FOLD, the resulting model of treatment developed for the delivery of the service. Setting As a partnership, we have come together to coproduce a model which allows us to share our clinical and industry knowledge and build on our reputations as trusted treatment providers. The presentation will outline our expertise share in development of a unified approach to recovery-oriented models of care, clinical governance, risk assessment and management and aftercare and continuous recovery. We will also introduce our innovative specialist referral portal which will offer referring partners the ability to include the service user in planning their own recovery journey. Outcomes Our collaboration has resulted in the development of the FOLD model which includes three agile and flexible treatment packages aimed at offering the most enhanced and comprehensive treatment in UK, to date, for those most affected by gambling harm. The paper will offer insight into each treatment package and all recovery model stages involved, as well as into the partnership work with NGST providers, local mental health and social care providers and lived experience organisation that will enable us to offer support to more 100 people a year who would otherwise get “lost in the system”. Conclusion FOLD offers a great opportunity to develop, implement and evaluate a new, much needed, whole-person and whole-system approach to counter gambling related harms.Keywords: gambling treatment, partnership working, integrated care pathways, NGTS, complex needs
Procedia PDF Downloads 134350 An Efficient Hardware/Software Workflow for Multi-Cores Simulink Applications
Authors: Asma Rebaya, Kaouther Gasmi, Imen Amari, Salem Hasnaoui
Abstract:
Over these last years, applications such as telecommunications, signal processing, digital communication with advanced features (Multi-antenna, equalization..) witness a rapid evaluation accompanied with an increase of user exigencies in terms of latency, the power of computation… To satisfy these requirements, the use of hardware/software systems is a common solution; where hardware is composed of multi-cores and software is represented by models of computation, synchronous data flow (SDF) graph for instance. Otherwise, the most of the embedded system designers utilize Simulink for modeling. The issue is how to simplify the c code generation, for a multi-cores platform, of an application modeled by Simulink. To overcome this problem, we propose a workflow allowing an automatic transformation from the Simulink model to the SDF graph and providing an efficient schedule permitting to optimize the number of cores and to minimize latency. This workflow goes from a Simulink application and a hardware architecture described by IP.XACT language. Based on the synchronous and hierarchical behavior of both models, the Simulink block diagram is automatically transformed into an SDF graph. Once this process is successfully achieved, the scheduler calculates the optimal cores’ number needful by minimizing the maximum density of the whole application. Then, a core is chosen to execute a specific graph task in a specific order and, subsequently, a compatible C code is generated. In order to perform this proposal, we extend Preesm, a rapid prototyping tool, to take the Simulink model as entry input and to support the optimal schedule. Afterward, we compared our results to this tool results, using a simple illustrative application. The comparison shows that our results strictly dominate the Preesm results in terms of number of cores and latency. In fact, if Preesm needs m processors and latency L, our workflow need processors and latency L'< L.Keywords: hardware/software system, latency, modeling, multi-cores platform, scheduler, SDF graph, Simulink model, workflow
Procedia PDF Downloads 268349 Disabilities in Railways: Proposed Changes to the Design of Railway Compartments for the Inclusion of Differently Abled Persons
Authors: Bathmajaa Muralisankar
Abstract:
As much as railway station infrastructure designs and ticket-booking norms have been changed to facilitate use by differently abled persons, the railway train compartments themselves have not been made user-friendly for differently abled persons. Owing to safety concerns, dependency on others for their travel, and fear of isolation, differently abled people do not prefer travelling by train. Rather than including a dedicated compartment open only to the differently abled, including the latter with others in the normal compartment (with the proposed modifications discussed here) will make them feel secure and make for an enhanced travel experience for them. This approach also represents the most practical way to include a particular category of people in the mainstream society. Lowering the height of the compartment doors and providing a wider entrance with a ramp will provide easy entry for those using wheelchairs. As well, removing the first two alternate rows and the first two side seats will not only widen the passage and increase seating space but also improve wheelchair turning radius. This will help them travel without having to depend on others. Seating arrangements may be done to accommodate their family members near them instead of isolating the differently abled in a separate compartment. According to present ticket-booking regulations of the Indian Railways, three to four disabled persons may travel without their family or one to two along with their family, and the numbers may be added or reduced. To help visually challenged and hearing-impaired persons, in addition to the provision of special instruments, railings, and textured footpaths and flooring, the seat numbers above the seats may be set in metal or plastic as an outward projection so the visually impaired can touch and feel the numbers. Braille boards may be included at the entrance to the compartment along with seat numbers in the aforementioned projected manner. These seat numbers may be designed as buttons, which when pressed results in an announcement of the seat number in the applicable local language as well as English. Emergency buttons, rather than emergency chains, within the easy reach of disabled passengers will also help them.Keywords: dependency, differently abled, inclusion, mainstream society
Procedia PDF Downloads 254348 Organ Dose Calculator for Fetus Undergoing Computed Tomography
Authors: Choonsik Lee, Les Folio
Abstract:
Pregnant patients may undergo CT in emergencies unrelated with pregnancy, and potential risk to the developing fetus is of concern. It is critical to accurately estimate fetal organ doses in CT scans. We developed a fetal organ dose calculation tool using pregnancy-specific computational phantoms combined with Monte Carlo radiation transport techniques. We adopted a series of pregnancy computational phantoms developed at the University of Florida at the gestational ages of 8, 10, 15, 20, 25, 30, 35, and 38 weeks (Maynard et al. 2011). More than 30 organs and tissues and 20 skeletal sites are defined in each fetus model. We calculated fetal organ dose-normalized by CTDIvol to derive organ dose conversion coefficients (mGy/mGy) for the eight fetuses for consequential slice locations ranging from the top to the bottom of the pregnancy phantoms with 1 cm slice thickness. Organ dose from helical scans was approximated by the summation of doses from multiple axial slices included in the given scan range of interest. We then compared dose conversion coefficients for major fetal organs in the abdominal-pelvis CT scan of pregnancy phantoms with the uterine dose of a non-pregnant adult female computational phantom. A comprehensive library of organ conversion coefficients was established for the eight developing fetuses undergoing CT. They were implemented into an in-house graphical user interface-based computer program for convenient estimation of fetal organ doses by inputting CT technical parameters as well as the age of the fetus. We found that the esophagus received the least dose, whereas the kidneys received the greatest dose in all fetuses in AP scans of the pregnancy phantoms. We also found that when the uterine dose of a non-pregnant adult female phantom is used as a surrogate for fetal organ doses, root-mean-square-error ranged from 0.08 mGy (8 weeks) to 0.38 mGy (38 weeks). The uterine dose was up to 1.7-fold greater than the esophagus dose of the 38-week fetus model. The calculation tool should be useful in cases requiring fetal organ dose in emergency CT scans as well as patient dose monitoring.Keywords: computed tomography, fetal dose, pregnant women, radiation dose
Procedia PDF Downloads 140347 An Advanced Automated Brain Tumor Diagnostics Approach
Authors: Berkan Ural, Arif Eser, Sinan Apaydin
Abstract:
Medical image processing is generally become a challenging task nowadays. Indeed, processing of brain MRI images is one of the difficult parts of this area. This study proposes a hybrid well-defined approach which is consisted from tumor detection, extraction and analyzing steps. This approach is mainly consisted from a computer aided diagnostics system for identifying and detecting the tumor formation in any region of the brain and this system is commonly used for early prediction of brain tumor using advanced image processing and probabilistic neural network methods, respectively. For this approach, generally, some advanced noise removal functions, image processing methods such as automatic segmentation and morphological operations are used to detect the brain tumor boundaries and to obtain the important feature parameters of the tumor region. All stages of the approach are done specifically with using MATLAB software. Generally, for this approach, firstly tumor is successfully detected and the tumor area is contoured with a specific colored circle by the computer aided diagnostics program. Then, the tumor is segmented and some morphological processes are achieved to increase the visibility of the tumor area. Moreover, while this process continues, the tumor area and important shape based features are also calculated. Finally, with using the probabilistic neural network method and with using some advanced classification steps, tumor area and the type of the tumor are clearly obtained. Also, the future aim of this study is to detect the severity of lesions through classes of brain tumor which is achieved through advanced multi classification and neural network stages and creating a user friendly environment using GUI in MATLAB. In the experimental part of the study, generally, 100 images are used to train the diagnostics system and 100 out of sample images are also used to test and to check the whole results. The preliminary results demonstrate the high classification accuracy for the neural network structure. Finally, according to the results, this situation also motivates us to extend this framework to detect and localize the tumors in the other organs.Keywords: image processing algorithms, magnetic resonance imaging, neural network, pattern recognition
Procedia PDF Downloads 418346 Eco-Friendly Softener Extracted from Ricinus communis (Castor) Seeds for Organic Cotton Fabric
Authors: Fisaha Asmelash
Abstract:
The processing of textiles to achieve a desired handle is a crucial aspect of finishing technology. Softeners can enhance the properties of textiles, such as softness, smoothness, elasticity, hydrophilicity, antistatic properties, and soil release properties, depending on the chemical nature used. However, human skin is sensitive to rough textiles, making softeners increasingly important. Although synthetic softeners are available, they are often expensive and can cause allergic reactions on human skin. This paper aims to extract a natural softener from Ricinus communis and produce an eco-friendly and user-friendly alternative due to its 100% herbal and organic nature. Crushed Ricinus communis seeds were soaked in a mechanical oil extractor for one hour with a 100g cotton fabric sample. The defatted cake or residue obtained after the extraction of oil from the seeds, also known as Ricinus communis meal, was obtained by filtering the raffinate and then dried at 1030c for four hours before being stored under laboratory conditions for the softening process. The softener was applied directly to 100% cotton fabric using the padding process, and the fabric was tested for stiffness, crease recovery, and drape ability. The effect of different concentrations of finishing agents on fabric stiffness, crease recovery, and drape ability was also analyzed. The results showed that the change in fabric softness depends on the concentration of the finish used. As the concentration of the finish was increased, there was a decrease in bending length and drape coefficient. Fabrics with a high concentration of softener showed a maximum decrease in drape coefficient and stiffness, comparable to commercial softeners such as silicon. The highest decrease in drape coefficient was found to be comparable with commercial softeners, silicon. Maximum increases in crease recovery were seen in fabrics treated with Ricinus communis softener at a concentration of 30gpl. From the results, the extracted softener proved to be effective in the treatment of 100% cotton fabricKeywords: ricinus communis, crease recovery, drapability, softeners, stiffness
Procedia PDF Downloads 91345 MIMO Radar-Based System for Structural Health Monitoring and Geophysical Applications
Authors: Davide D’Aria, Paolo Falcone, Luigi Maggi, Aldo Cero, Giovanni Amoroso
Abstract:
The paper presents a methodology for real-time structural health monitoring and geophysical applications. The key elements of the system are a high performance MIMO RADAR sensor, an optical camera and a dedicated set of software algorithms encompassing interferometry, tomography and photogrammetry. The MIMO Radar sensor proposed in this work, provides an extremely high sensitivity to displacements making the system able to react to tiny deformations (up to tens of microns) with a time scale which spans from milliseconds to hours. The MIMO feature of the system makes the system capable of providing a set of two-dimensional images of the observed scene, each mapped on the azimuth-range directions with noticeably resolution in both the dimensions and with an outstanding repetition rate. The back-scattered energy, which is distributed in the 3D space, is projected on a 2D plane, where each pixel has as coordinates the Line-Of-Sight distance and the cross-range azimuthal angle. At the same time, the high performing processing unit allows to sense the observed scene with remarkable refresh periods (up to milliseconds), thus opening the way for combined static and dynamic structural health monitoring. Thanks to the smart TX/RX antenna array layout, the MIMO data can be processed through a tomographic approach to reconstruct the three-dimensional map of the observed scene. This 3D point cloud is then accurately mapped on a 2D digital optical image through photogrammetric techniques, allowing for easy and straightforward interpretations of the measurements. Once the three-dimensional image is reconstructed, a 'repeat-pass' interferometric approach is exploited to provide the user of the system with high frequency three-dimensional motion/vibration estimation of each point of the reconstructed image. At this stage, the methodology leverages consolidated atmospheric correction algorithms to provide reliable displacement and vibration measurements.Keywords: interferometry, MIMO RADAR, SAR, tomography
Procedia PDF Downloads 195344 Forecasting Future Society to Explore Promising Security Technologies
Authors: Jeonghwan Jeon, Mintak Han, Youngjun Kim
Abstract:
Due to the rapid development of information and communication technology (ICT), a substantial transformation is currently happening in the society. As the range of intelligent technologies and services is continuously expanding, ‘things’ are becoming capable of communicating one another and even with people. However, such “Internet of Things” has the technical weakness so that a great amount of such information transferred in real-time may be widely exposed to the threat of security. User’s personal data are a typical example which is faced with a serious security threat. The threats of security will be diversified and arose more frequently because next generation of unfamiliar technology develops. Moreover, as the society is becoming increasingly complex, security vulnerability will be increased as well. In the existing literature, a considerable number of private and public reports that forecast future society have been published as a precedent step of the selection of future technology and the establishment of strategies for competitiveness. Although there are previous studies that forecast security technology, they have focused only on technical issues and overlooked the interrelationships between security technology and social factors are. Therefore, investigations of security threats in the future and security technology that is able to protect people from various threats are required. In response, this study aims to derive potential security threats associated with the development of technology and to explore the security technology that can protect against them. To do this, first of all, private and public reports that forecast future and online documents from technology-related communities are collected. By analyzing the data, future issues are extracted and categorized in terms of STEEP (Society, Technology, Economy, Environment, and Politics), as well as security. Second, the components of potential security threats are developed based on classified future issues. Then, points that the security threats may occur –for example, mobile payment system based on a finger scan technology– are identified. Lastly, alternatives that prevent potential security threats are proposed by matching security threats with points and investigating related security technologies from patent data. Proposed approach can identify the ICT-related latent security menaces and provide the guidelines in the ‘problem – alternative’ form by linking the threat point with security technologies.Keywords: future society, information and communication technology, security technology, technology forecasting
Procedia PDF Downloads 468343 Impact of the Non-Energy Sectors Diversification on the Energy Dependency Mitigation: Visualization by the “IntelSymb” Software Application
Authors: Ilaha Rzayeva, Emin Alasgarov, Orkhan Karim-Zada
Abstract:
This study attempts to consider the linkage between management and computer sciences in order to develop the software named “IntelSymb” as a demo application to prove data analysis of non-energy* fields’ diversification, which will positively influence on energy dependency mitigation of countries. Afterward, we analyzed 18 years of economic fields of development (5 sectors) of 13 countries by identifying which patterns mostly prevailed and which can be dominant in the near future. To make our analysis solid and plausible, as a future work, we suggest developing a gateway or interface, which will be connected to all available on-line data bases (WB, UN, OECD, U.S. EIA) for countries’ analysis by fields. Sample data consists of energy (TPES and energy import indicators) and non-energy industries’ (Main Science and Technology Indicator, Internet user index, and Sales and Production indicators) statistics from 13 OECD countries over 18 years (1995-2012). Our results show that the diversification of non-energy industries can have a positive effect on energy sector dependency (energy consumption and import dependence on crude oil) deceleration. These results can provide empirical and practical support for energy and non-energy industries diversification’ policies, such as the promoting of Information and Communication Technologies (ICTs), services and innovative technologies efficiency and management, in other OECD and non-OECD member states with similar energy utilization patterns and policies. Industries, including the ICT sector, generate around 4 percent of total GHG, but this is much higher — around 14 percent — if indirect energy use is included. The ICT sector itself (excluding the broadcasting sector) contributes approximately 2 percent of global GHG emissions, at just under 1 gigatonne of carbon dioxide equivalent (GtCO2eq). Ergo, this can be a good example and lesson for countries which are dependent and independent on energy, and mainly emerging oil-based economies, as well as to motivate non-energy industries diversification in order to be ready to energy crisis and to be able to face any economic crisis as well.Keywords: energy policy, energy diversification, “IntelSymb” software, renewable energy
Procedia PDF Downloads 224342 Multi-Agent System Based Solution for Operating Agile and Customizable Micro Manufacturing Systems
Authors: Dylan Santos De Pinho, Arnaud Gay De Combes, Matthieu Steuhlet, Claude Jeannerat, Nabil Ouerhani
Abstract:
The Industry 4.0 initiative has been launched to address huge challenges related to ever-smaller batch sizes. The end-user need for highly customized products requires highly adaptive production systems in order to keep the same efficiency of shop floors. Most of the classical Software solutions that operate the manufacturing processes in a shop floor are based on rigid Manufacturing Execution Systems (MES), which are not capable to adapt the production order on the fly depending on changing demands and or conditions. In this paper, we present a highly modular and flexible solution to orchestrate a set of production systems composed of a micro-milling machine-tool, a polishing station, a cleaning station, a part inspection station, and a rough material store. The different stations are installed according to a novel matrix configuration of a 3x3 vertical shelf. The different cells of the shelf are connected through horizontal and vertical rails on which a set of shuttles circulate to transport the machined parts from a station to another. Our software solution for orchestrating the tasks of each station is based on a Multi-Agent System. Each station and each shuttle is operated by an autonomous agent. All agents communicate with a central agent that holds all the information about the manufacturing order. The core innovation of this paper lies in the path planning of the different shuttles with two major objectives: 1) reduce the waiting time of stations and thus reduce the cycle time of the entire part, and 2) reduce the disturbances like vibration generated by the shuttles, which highly impacts the manufacturing process and thus the quality of the final part. Simulation results show that the cycle time of the parts is reduced by up to 50% compared with MES operated linear production lines while the disturbance is systematically avoided for the critical stations like the milling machine-tool.Keywords: multi-agent systems, micro-manufacturing, flexible manufacturing, transfer systems
Procedia PDF Downloads 130341 Using Business Intelligence Capabilities to Improve the Quality of Decision-Making: A Case Study of Mellat Bank
Authors: Jalal Haghighat Monfared, Zahra Akbari
Abstract:
Today, business executives need to have useful information to make better decisions. Banks have also been using information tools so that they can direct the decision-making process in order to achieve their desired goals by rapidly extracting information from sources with the help of business intelligence. The research seeks to investigate whether there is a relationship between the quality of decision making and the business intelligence capabilities of Mellat Bank. Each of the factors studied is divided into several components, and these and their relationships are measured by a questionnaire. The statistical population of this study consists of all managers and experts of Mellat Bank's General Departments (including 190 people) who use commercial intelligence reports. The sample size of this study was 123 randomly determined by statistical method. In this research, relevant statistical inference has been used for data analysis and hypothesis testing. In the first stage, using the Kolmogorov-Smirnov test, the normalization of the data was investigated and in the next stage, the construct validity of both variables and their resulting indexes were verified using confirmatory factor analysis. Finally, using the structural equation modeling and Pearson's correlation coefficient, the research hypotheses were tested. The results confirmed the existence of a positive relationship between decision quality and business intelligence capabilities in Mellat Bank. Among the various capabilities, including data quality, correlation with other systems, user access, flexibility and risk management support, the flexibility of the business intelligence system was the most correlated with the dependent variable of the present research. This shows that it is necessary for Mellat Bank to pay more attention to choose the required business intelligence systems with high flexibility in terms of the ability to submit custom formatted reports. Subsequently, the quality of data on business intelligence systems showed the strongest relationship with quality of decision making. Therefore, improving the quality of data, including the source of data internally or externally, the type of data in quantitative or qualitative terms, the credibility of the data and perceptions of who uses the business intelligence system, improves the quality of decision making in Mellat Bank.Keywords: business intelligence, business intelligence capability, decision making, decision quality
Procedia PDF Downloads 112