Search results for: machine intelligence
889 Diagnosis and Analysis of Automated Liver and Tumor Segmentation on CT
Authors: R. R. Ramsheeja, R. Sreeraj
Abstract:
For view the internal structures of the human body such as liver, brain, kidney etc have a wide range of different modalities for medical images are provided nowadays. Computer Tomography is one of the most significant medical image modalities. In this paper use CT liver images for study the use of automatic computer aided techniques to calculate the volume of the liver tumor. Segmentation method is used for the detection of tumor from the CT scan is proposed. Gaussian filter is used for denoising the liver image and Adaptive Thresholding algorithm is used for segmentation. Multiple Region Of Interest(ROI) based method that may help to characteristic the feature different. It provides a significant impact on classification performance. Due to the characteristic of liver tumor lesion, inherent difficulties appear selective. For a better performance, a novel proposed system is introduced. Multiple ROI based feature selection and classification are performed. In order to obtain of relevant features for Support Vector Machine(SVM) classifier is important for better generalization performance. The proposed system helps to improve the better classification performance, reason in which we can see a significant reduction of features is used. The diagnosis of liver cancer from the computer tomography images is very difficult in nature. Early detection of liver tumor is very helpful to save the human life.Keywords: computed tomography (CT), multiple region of interest(ROI), feature values, segmentation, SVM classification
Procedia PDF Downloads 508888 Machining Responce of Austempered Ductile Iron with Varying Cutting Speed and Depth of Cut
Authors: Prashant Parhad, Vinayak Dakre, Ajay Likhite, Jatin Bhatt
Abstract:
This work mainly focuses on machinability studies of Austempered Ductile Iron (ADI). The Ductile Iron (DI) was austempered at 250 oC for different durations and the process window for austempering was established by studying the microstructure. The microstructural characterization of the material was done using optical microscopy, SEM and XRD. The samples austempered as per the process window were then subjected to turning using a TiAlN-coated tungsten carbide insert to study the effect of cutting parameters, namely the cutting speed and the depth of cut. The effect was investigated in terms of cutting forces required as well as the surface roughness obtained. The turning was conducted on a CNC turning machine and primary (Fx), radial (Fy) and feed (Fz) cutting forces were quantified with a three-component dynamometer. It was observed that the magnitude of radial force was more than that of primary cutting force for all cutting speed and for various depths of cut studied. It has also been seen that increasing the cutting speed improves the surface quality. The observed machinability behaviour was investigated in light of the microstructure of the material obtained under the given austempering conditions and a structure-property- co-relation was established between the two. For all cutting speed and depth of cut, the best machining response in terms of cutting forces and surface quality was obtained towards the centre of process window.Keywords: process window, cutting speed, depth of cut, surface roughness
Procedia PDF Downloads 367887 Prompt Design for Code Generation in Data Analysis Using Large Language Models
Authors: Lu Song Ma Li Zhi
Abstract:
With the rapid advancement of artificial intelligence technology, large language models (LLMs) have become a milestone in the field of natural language processing, demonstrating remarkable capabilities in semantic understanding, intelligent question answering, and text generation. These models are gradually penetrating various industries, particularly showcasing significant application potential in the data analysis domain. However, retraining or fine-tuning these models requires substantial computational resources and ample downstream task datasets, which poses a significant challenge for many enterprises and research institutions. Without modifying the internal parameters of the large models, prompt engineering techniques can rapidly adapt these models to new domains. This paper proposes a prompt design strategy aimed at leveraging the capabilities of large language models to automate the generation of data analysis code. By carefully designing prompts, data analysis requirements can be described in natural language, which the large language model can then understand and convert into executable data analysis code, thereby greatly enhancing the efficiency and convenience of data analysis. This strategy not only lowers the threshold for using large models but also significantly improves the accuracy and efficiency of data analysis. Our approach includes requirements for the precision of natural language descriptions, coverage of diverse data analysis needs, and mechanisms for immediate feedback and adjustment. Experimental results show that with this prompt design strategy, large language models perform exceptionally well in multiple data analysis tasks, generating high-quality code and significantly shortening the data analysis cycle. This method provides an efficient and convenient tool for the data analysis field and demonstrates the enormous potential of large language models in practical applications.Keywords: large language models, prompt design, data analysis, code generation
Procedia PDF Downloads 37886 Eliminating Arm, Neck and Leg Fatigue of United Asia International Plastics Corporation Workers through Rapid Entire Body Assessment
Authors: John Cheferson R. De Belen, John Paul G. Elizares, Ronald John G. Raz, Janina Elyse A. Reyes, Charie G. Salengua, Aristotle L. Soriano
Abstract:
Plastic is a type of synthetic or man-made polymer that can readily be molded into a variety of products. Its usage over the past century has enabled society to make huge technological advances. The workers of United Asia International Plastics Corporation (UAIPC), a plastic manufacturing company performs manual packaging which causes fatigue and stress on their arm, neck, and legs due to extended periods of standing and repetitive motions. With the use of the Fishbone Diagram, Five-Why Analysis, Rapid Entire Body Assessment (REBA), and Anthropometry, the stressful tasks and activities were identified and analyzed. Given the anthropometric measurements obtained from the workers, improved dimensions for the tables and chairs should be used and provide a new packaging machine. The validation of this proposal shall follow after its implementation. By eliminating fatigue during working hours in the production, the workers will be at ease at performing their work properly; productivity will increase that will lead to more profit. Further areas for study include measurement and comparison of the worker’s anthropometric measurement with the industry standard.Keywords: anthropometry, fishbone diagram, five-why analysis, rapid entire body assessment
Procedia PDF Downloads 263885 Developing an ANN Model to Predict Anthropometric Dimensions Based on Real Anthropometric Database
Authors: Waleed A. Basuliman, Khalid S. AlSaleh, Mohamed Z. Ramadan
Abstract:
Applying the anthropometric dimensions is considered one of the important factors when designing any human-machine system. In this study, the estimation of anthropometric dimensions has been improved by developing artificial neural network that aims to predict the anthropometric measurements of the male in Saudi Arabia. A total of 1427 Saudi males from age 6 to 60 participated in measuring twenty anthropometric dimensions. These anthropometric measurements are important for designing the majority of work and life applications in Saudi Arabia. The data were collected during 8 months from different locations in Riyadh City. Five of these dimensions were used as predictors variables (inputs) of the model, and the remaining fifteen dimensions were set to be the measured variables (outcomes). The hidden layers have been varied during the structuring stage, and the best performance was achieved with the network structure 6-25-15. The results showed that the developed Neural Network model was significantly able to predict the body dimensions for the population of Saudi Arabia. The network mean absolute percentage error (MAPE) and the root mean squared error (RMSE) were found 0.0348 and 3.225 respectively. The accuracy of the developed neural network was evaluated by compare the predicted outcomes with a multiple regression model. The ANN model performed better and resulted excellent correlation coefficients between the predicted and actual dimensions.Keywords: artificial neural network, anthropometric measurements, backpropagation, real anthropometric database
Procedia PDF Downloads 574884 A Mechanical Diagnosis Method Based on Vibration Fault Signal down-Sampling and the Improved One-Dimensional Convolutional Neural Network
Authors: Bowei Yuan, Shi Li, Liuyang Song, Huaqing Wang, Lingli Cui
Abstract:
Convolutional neural networks (CNN) have received extensive attention in the field of fault diagnosis. Many fault diagnosis methods use CNN for fault type identification. However, when the amount of raw data collected by sensors is massive, the neural network needs to perform a time-consuming classification task. In this paper, a mechanical fault diagnosis method based on vibration signal down-sampling and the improved one-dimensional convolutional neural network is proposed. Through the robust principal component analysis, the low-rank feature matrix of a large amount of raw data can be separated, and then down-sampling is realized to reduce the subsequent calculation amount. In the improved one-dimensional CNN, a smaller convolution kernel is used to reduce the number of parameters and computational complexity, and regularization is introduced before the fully connected layer to prevent overfitting. In addition, the multi-connected layers can better generalize classification results without cumbersome parameter adjustments. The effectiveness of the method is verified by monitoring the signal of the centrifugal pump test bench, and the average test accuracy is above 98%. When compared with the traditional deep belief network (DBN) and support vector machine (SVM) methods, this method has better performance.Keywords: fault diagnosis, vibration signal down-sampling, 1D-CNN
Procedia PDF Downloads 130883 Breast Cancer Risk is Predicted Using Fuzzy Logic in MATLAB Environment
Authors: S. Valarmathi, P. B. Harathi, R. Sridhar, S. Balasubramanian
Abstract:
Machine learning tools in medical diagnosis is increasing due to the improved effectiveness of classification and recognition systems to help medical experts in diagnosing breast cancer. In this study, ID3 chooses the splitting attribute with the highest gain in information, where gain is defined as the difference between before the split versus after the split. It is applied for age, location, taluk, stage, year, period, martial status, treatment, heredity, sex, and habitat against Very Serious (VS), Very Serious Moderate (VSM), Serious (S) and Not Serious (NS) to calculate the gain of information. The ranked histogram gives the gain of each field for the breast cancer data. The doctors use TNM staging which will decide the risk level of the breast cancer and play an important decision making field in fuzzy logic for perception based measurement. Spatial risk area (taluk) of the breast cancer is calculated. Result clearly states that Coimbatore (North and South) was found to be risk region to the breast cancer than other areas at 20% criteria. Weighted value of taluk was compared with criterion value and integrated with Map Object to visualize the results. ID3 algorithm shows the high breast cancer risk regions in the study area. The study has outlined, discussed and resolved the algorithms, techniques / methods adopted through soft computing methodology like ID3 algorithm for prognostic decision making in the seriousness of the breast cancer.Keywords: ID3 algorithm, breast cancer, fuzzy logic, MATLAB
Procedia PDF Downloads 516882 Equation for Predicting Inferior Vena Cava Diameter as a Potential Pointer for Heart Failure Diagnosis among Adult in Azare, Bauchi State, Nigeria
Authors: M. K. Yusuf, W. O. Hamman, U. E. Umana, S. B. Oladele
Abstract:
Background: Dilatation of the inferior vena cava (IVC) is used as the ultrasonic diagnostic feature in patients suspected of congestive heart failure. The IVC diameter has been reported to vary among the various body mass indexes (BMI) and body shape indexes (ABSI). Knowledge of these variations is useful in precision diagnoses of CHF by imaging scientists. Aim: The study aimed to establish an equation for predicting the ultrasonic mean diameter of the IVC among the various BMI/ABSI of inhabitants of Azare, Bauchi State-Nigeria. Methodology: Two hundred physically healthy adult subjects of both sexes were classified into under, normal, over, and obese weights using their BMIs after selection using a structured questionnaire following their informed consent for an abdominal ultrasound scan. The probe was placed on the midline of the body, halfway between the xiphoid process and the umbilicus, with the marker on the probe directed towards the patient's head to obtain a longitudinal view of the IVC. The maximum IVC diameter was measured from the subcostal view using the electronic caliper of the scan machine. The mean value of each group was obtained, and the results were analysed. Results: A novel equation {(IVC Diameter = 1.04 +0.01(X) where X= BMI} has been generated for determining the IVC diameter among the populace. Conclusion: An equation for predicting the IVC diameter from individual BMI values in apparently healthy subjects has been established.Keywords: equation, ultrasonic, IVC diameter, body adiposities
Procedia PDF Downloads 68881 A Data-Mining Model for Protection of FACTS-Based Transmission Line
Authors: Ashok Kalagura
Abstract:
This paper presents a data-mining model for fault-zone identification of flexible AC transmission systems (FACTS)-based transmission line including a thyristor-controlled series compensator (TCSC) and unified power-flow controller (UPFC), using ensemble decision trees. Given the randomness in the ensemble of decision trees stacked inside the random forests model, it provides an effective decision on the fault-zone identification. Half-cycle post-fault current and voltage samples from the fault inception are used as an input vector against target output ‘1’ for the fault after TCSC/UPFC and ‘1’ for the fault before TCSC/UPFC for fault-zone identification. The algorithm is tested on simulated fault data with wide variations in operating parameters of the power system network, including noisy environment providing a reliability measure of 99% with faster response time (3/4th cycle from fault inception). The results of the presented approach using the RF model indicate the reliable identification of the fault zone in FACTS-based transmission lines.Keywords: distance relaying, fault-zone identification, random forests, RFs, support vector machine, SVM, thyristor-controlled series compensator, TCSC, unified power-flow controller, UPFC
Procedia PDF Downloads 421880 A Kernel-Based Method for MicroRNA Precursor Identification
Authors: Bin Liu
Abstract:
MicroRNAs (miRNAs) are small non-coding RNA molecules, functioning in transcriptional and post-transcriptional regulation of gene expression. The discrimination of the real pre-miRNAs from the false ones (such as hairpin sequences with similar stem-loops) is necessary for the understanding of miRNAs’ role in the control of cell life and death. Since both their small size and sequence specificity, it cannot be based on sequence information alone but requires structure information about the miRNA precursor to get satisfactory performance. Kmers are convenient and widely used features for modeling the properties of miRNAs and other biological sequences. However, Kmers suffer from the inherent limitation that if the parameter K is increased to incorporate long range effects, some certain Kmer will appear rarely or even not appear, as a consequence, most Kmers absent and a few present once. Thus, the statistical learning approaches using Kmers as features become susceptible to noisy data once K becomes large. In this study, we proposed a Gapped k-mer approach to overcome the disadvantages of Kmers, and applied this method to the field of miRNA prediction. Combined with the structure status composition, a classifier called imiRNA-GSSC was proposed. We show that compared to the original imiRNA-kmer and alternative approaches. Trained on human miRNA precursors, this predictor can achieve an accuracy of 82.34 for predicting 4022 pre-miRNA precursors from eleven species.Keywords: gapped k-mer, imiRNA-GSSC, microRNA precursor, support vector machine
Procedia PDF Downloads 159879 The Impact of Artificial Intelligence on Pharmacy and Pharmacology
Authors: Mamdouh Milad Adly Morkos
Abstract:
Despite having the greatest rates of mortality and morbidity in the world, low- and middle-income (LMIC) nations trail high-income nations in terms of the number of clinical trials, the number of qualified researchers, and the amount of research information specific to their people. Health inequities and the use of precision medicine may be hampered by a lack of local genomic data, clinical pharmacology and pharmacometrics competence, and training opportunities. These issues can be solved by carrying out health care infrastructure development, which includes data gathering and well-designed clinical pharmacology training in LMICs. It will be advantageous if there is international cooperation focused at enhancing education and infrastructure and promoting locally motivated clinical trials and research. This paper outlines various instances where clinical pharmacology knowledge could be put to use, including pharmacogenomic opportunities that could lead to better clinical guideline recommendations. Examples of how clinical pharmacology training can be successfully implemented in LMICs are also provided, including clinical pharmacology and pharmacometrics training programmes in Africa and a Tanzanian researcher's personal experience while on a training sabbatical in the United States. These training initiatives will profit from advocacy for clinical pharmacologists' employment prospects and career development pathways, which are gradually becoming acknowledged and established in LMICs. The advancement of training and research infrastructure to increase clinical pharmacologists' knowledge in LMICs would be extremely beneficial because they have a significant role to play in global healthKeywords: electromagnetic solar system, nano-material, nano pharmacology, pharmacovigilance, quantum theoryclinical simulation, education, pharmacology, simulation, virtual learning low- and middle-income, clinical pharmacology, pharmacometrics, career development pathways
Procedia PDF Downloads 81878 A Methodology to Integrate Data in the Company Based on the Semantic Standard in the Context of Industry 4.0
Authors: Chang Qin, Daham Mustafa, Abderrahmane Khiat, Pierre Bienert, Paulo Zanini
Abstract:
Nowadays, companies are facing lots of challenges in the process of digital transformation, which can be a complex and costly undertaking. Digital transformation involves the collection and analysis of large amounts of data, which can create challenges around data management and governance. Furthermore, it is also challenged to integrate data from multiple systems and technologies. Although with these pains, companies are still pursuing digitalization because by embracing advanced technologies, companies can improve efficiency, quality, decision-making, and customer experience while also creating different business models and revenue streams. In this paper, the issue that data is stored in data silos with different schema and structures is focused. The conventional approaches to addressing this issue involve utilizing data warehousing, data integration tools, data standardization, and business intelligence tools. However, these approaches primarily focus on the grammar and structure of the data and neglect the importance of semantic modeling and semantic standardization, which are essential for achieving data interoperability. In this session, the challenge of data silos in Industry 4.0 is addressed by developing a semantic modeling approach compliant with Asset Administration Shell (AAS) models as an efficient standard for communication in Industry 4.0. The paper highlights how our approach can facilitate the data mapping process and semantic lifting according to existing industry standards such as ECLASS and other industrial dictionaries. It also incorporates the Asset Administration Shell technology to model and map the company’s data and utilize a knowledge graph for data storage and exploration.Keywords: data interoperability in industry 4.0, digital integration, industrial dictionary, semantic modeling
Procedia PDF Downloads 93877 Railway Crane Accident: A Comparative Metallographic Test on Pins Fractured during Operation
Authors: Thiago Viana
Abstract:
Eventually train accidents occur on railways and for some specific cases it is necessary to use a train rescue with a crane positioned under a platform wagon. These tumbled machines are collected and sent to the machine shop or scrap yard. In one of these cranes that were being used to rescue a wagon, occurred a fall of hoist due to fracture of two large pins. The two pins were collected and sent for failure analysis. This work investigates the main cause and the secondary causes for the initiation of the fatigue crack. All standard failure analysis procedures were applied, with careful evaluation of the characteristics of the material, fractured surfaces and, mainly, metallographic tests using an optical microscope to compare the geometry of the peaks and valleys of the thread of the pins and their respective seats. By metallographic analysis, it was concluded that the fatigue cracks were started from a notch (stress concentration) in the valley of the threads of the pin applied to the right side of the crane (pin 1). In this, it was verified that the peaks of the threads of the pin seat did not have proper geometry, with sharp edges being present that caused such notches. The visual analysis showed that fracture of the pin on the left side of the crane (pin 2) was brittle type, being a consequence of the fracture of the first one. Recommendations for this and other railway cranes have been made, such as nondestructive testing, stress calculation, design review, quality control and suitability of the mechanical forming process of the seat threads and pin threads.Keywords: crane, fracture, pin, railway
Procedia PDF Downloads 106876 Determining Optimal Number of Trees in Random Forests
Authors: Songul Cinaroglu
Abstract:
Background: Random Forest is an efficient, multi-class machine learning method using for classification, regression and other tasks. This method is operating by constructing each tree using different bootstrap sample of the data. Determining the number of trees in random forests is an open question in the literature for studies about improving classification performance of random forests. Aim: The aim of this study is to analyze whether there is an optimal number of trees in Random Forests and how performance of Random Forests differ according to increase in number of trees using sample health data sets in R programme. Method: In this study we analyzed the performance of Random Forests as the number of trees grows and doubling the number of trees at every iteration using “random forest” package in R programme. For determining minimum and optimal number of trees we performed Mc Nemar test and Area Under ROC Curve respectively. Results: At the end of the analysis it was found that as the number of trees grows, it does not always means that the performance of the forest is better than forests which have fever trees. In other words larger number of trees only increases computational costs but not increases performance results. Conclusion: Despite general practice in using random forests is to generate large number of trees for having high performance results, this study shows that increasing number of trees doesn’t always improves performance. Future studies can compare different kinds of data sets and different performance measures to test whether Random Forest performance results change as number of trees increase or not.Keywords: classification methods, decision trees, number of trees, random forest
Procedia PDF Downloads 394875 A Study on the Effects of a Mindfulness Training on Managers: The Case of the Malian Company for the Development of Textile
Authors: Aboubacar Garba Konte, Wei Jun, Li Xiaohui
Abstract:
Nowadays companies are facing increasing pressure. The market environment changes more frequently than ever. Therefore, managers have to develop their agility, their performance and their capacity for innovation. Most companies look for managerial innovations to develop in their employees qualities such as motivation, commitment, creativity, autonomy or even the ability to adapt to change and manage intensive pressure. On a more collective level, companies are looking for teams that are able to organize, communicate and develop a form of collective intelligence based on cooperation and solidarity. Among the many managerial innovations that are currently developing, mindfulness (or mindfulness) is drawing the attention of a growing number of companies (Google, Apple, Sony, ING ...), These companies have implemented programs based on mindfulness. Although the concept of mindfulness and its effects have been the subject of in-depth research in the psychological field, research on mindfulness in the field of management is still in its infancy and it is necessary to evaluate its contribution to organizations. The purpose of this research is to evaluate the effects of a mindfulness training among the managers of a Malian textile company (CMDT). We conducted a case study on their experience and their managerial practices. In addition, we discuss the innovative nature of mindfulness in terms of managerial practice The results show significant positive effects on two major skills identified by managers that raise significant difficulties in their daily lives: their ability to supervise a team of employees with all that this implies in terms of interpersonal skills and their ability to organize and prioritize their activities. In addition, the research methodology sheds light on the innovative nature of mindfulness in a favorable organizational environment.Keywords: mindfulness, manager, managerial innovation, relational skills, organization and prioritization
Procedia PDF Downloads 100874 Computer-Aided Classification of Liver Lesions Using Contrasting Features Difference
Authors: Hussein Alahmer, Amr Ahmed
Abstract:
Liver cancer is one of the common diseases that cause the death. Early detection is important to diagnose and reduce the incidence of death. Improvements in medical imaging and image processing techniques have significantly enhanced interpretation of medical images. Computer-Aided Diagnosis (CAD) systems based on these techniques play a vital role in the early detection of liver disease and hence reduce liver cancer death rate. This paper presents an automated CAD system consists of three stages; firstly, automatic liver segmentation and lesion’s detection. Secondly, extracting features. Finally, classifying liver lesions into benign and malignant by using the novel contrasting feature-difference approach. Several types of intensity, texture features are extracted from both; the lesion area and its surrounding normal liver tissue. The difference between the features of both areas is then used as the new lesion descriptors. Machine learning classifiers are then trained on the new descriptors to automatically classify liver lesions into benign or malignant. The experimental results show promising improvements. Moreover, the proposed approach can overcome the problems of varying ranges of intensity and textures between patients, demographics, and imaging devices and settings.Keywords: CAD system, difference of feature, fuzzy c means, lesion detection, liver segmentation
Procedia PDF Downloads 323873 Comparison of Whole-Body Vibration and Plyometric Exercises on Explosive Power in Non-Athlete Girl Students
Authors: Fereshteh Zarei, Mahdi Kohandel
Abstract:
The aim of this study was investigate and compare plyometric and vibration exercises on muscle explosive power in non-athlete female students. For this purpose, 45 female students from non-athletes selected target then divided in to the three groups, two experimental and one control groups. From all groups were getting pre-tested. Experimental A did whole-body vibration exercises involved standing on one of machine vibration with frequency 30 Hz, amplitude 10 mm and in 5 different postures. Training for each position was 40 seconds with 60 seconds rest between it, and each season 5 seconds was added to duration of each body condition, until time up to 2 minutes for each postures. Exercises were done three times a week for 2 month. Experimental group B did plyometric exercises that include jumping, such as horizontal, vertical, and skipping .They included 10 times repeat for 5 set in each season. Intensity with increasing repetitions and sets were added. At this time, asked from control group that keep a daily activity and avoided strength training, explosive power and. after do exercises by groups we measured factors again. One-way analysis of variance and paired t statistical methods were used to analyze the data. There was significant difference in the amount of explosive power between the control and vibration groups (p=0/048) there was significant difference between the control and plyometric groups (019/0 = p). But between vibration and plyometric groups didn't observe significant difference in the amount of explosive power.Keywords: vibration, plyometric, exercises, explosive power, non-athlete
Procedia PDF Downloads 451872 Solving a Micromouse Maze Using an Ant-Inspired Algorithm
Authors: Rolando Barradas, Salviano Soares, António Valente, José Alberto Lencastre, Paulo Oliveira
Abstract:
This article reviews the Ant Colony Optimization, a nature-inspired algorithm, and its implementation in the Scratch/m-Block programming environment. The Ant Colony Optimization is a part of Swarm Intelligence-based algorithms and is a subset of biological-inspired algorithms. Starting with a problem in which one has a maze and needs to find its path to the center and return to the starting position. This is similar to an ant looking for a path to a food source and returning to its nest. Starting with the implementation of a simple wall follower simulator, the proposed solution uses a dynamic graphical interface that allows young students to observe the ants’ movement while the algorithm optimizes the routes to the maze’s center. Things like interface usability, Data structures, and the conversion of algorithmic language to Scratch syntax were some of the details addressed during this implementation. This gives young students an easier way to understand the computational concepts of sequences, loops, parallelism, data, events, and conditionals, as they are used through all the implemented algorithms. Future work includes the simulation results with real contest mazes and two different pheromone update methods and the comparison with the optimized results of the winners of each one of the editions of the contest. It will also include the creation of a Digital Twin relating the virtual simulator with a real micromouse in a full-size maze. The first test results show that the algorithm found the same optimized solutions that were found by the winners of each one of the editions of the Micromouse contest making this a good solution for maze pathfinding.Keywords: nature inspired algorithms, scratch, micromouse, problem-solving, computational thinking
Procedia PDF Downloads 125871 Potentials of Additive Manufacturing: An Approach to Increase the Flexibility of Production Systems
Authors: A. Luft, S. Bremen, N. Balc
Abstract:
The task of flexibility planning and design, just like factory planning, for example, is to create the long-term systemic framework that constitutes the restriction for short-term operational management. This is a strategic challenge since, due to the decision defect character of the underlying flexibility problem, multiple types of flexibility need to be considered over the course of various scenarios, production programs, and production system configurations. In this context, an evaluation model has been developed that integrates both conventional and additive resources on a basic task level and allows the quantification of flexibility enhancement in terms of mix and volume flexibility, complexity reduction, and machine capacity. The model helps companies to decide in early decision-making processes about the potential gains of implementing additive manufacturing technologies on a strategic level. For companies, it is essential to consider both additive and conventional manufacturing beyond pure unit costs. It is necessary to achieve an integrative view of manufacturing that incorporates both additive and conventional manufacturing resources and quantifies their potential with regard to flexibility and manufacturing complexity. This also requires a structured process for the strategic production systems design that spans the design of various scenarios and allows for multi-dimensional and comparative analysis. A respective guideline for the planning of additive resources on a strategic level is being laid out in this paper.Keywords: additive manufacturing, production system design, flexibility enhancement, strategic guideline
Procedia PDF Downloads 122870 Deep Learning-Based Approach to Automatic Abstractive Summarization of Patent Documents
Authors: Sakshi V. Tantak, Vishap K. Malik, Neelanjney Pilarisetty
Abstract:
A patent is an exclusive right granted for an invention. It can be a product or a process that provides an innovative method of doing something, or offers a new technical perspective or solution to a problem. A patent can be obtained by making the technical information and details about the invention publicly available. The patent owner has exclusive rights to prevent or stop anyone from using the patented invention for commercial uses. Any commercial usage, distribution, import or export of a patented invention or product requires the patent owner’s consent. It has been observed that the central and important parts of patents are scripted in idiosyncratic and complex linguistic structures that can be difficult to read, comprehend or interpret for the masses. The abstracts of these patents tend to obfuscate the precise nature of the patent instead of clarifying it via direct and simple linguistic constructs. This makes it necessary to have an efficient access to this knowledge via concise and transparent summaries. However, as mentioned above, due to complex and repetitive linguistic constructs and extremely long sentences, common extraction-oriented automatic text summarization methods should not be expected to show a remarkable performance when applied to patent documents. Other, more content-oriented or abstractive summarization techniques are able to perform much better and generate more concise summaries. This paper proposes an efficient summarization system for patents using artificial intelligence, natural language processing and deep learning techniques to condense the knowledge and essential information from a patent document into a single summary that is easier to understand without any redundant formatting and difficult jargon.Keywords: abstractive summarization, deep learning, natural language Processing, patent document
Procedia PDF Downloads 120869 Influence of Displacement Amplitude and Vertical Load on the Horizontal Dynamic and Static Behavior of Helical Wire Rope Isolators
Authors: Nicolò Vaiana, Mariacristina Spizzuoco, Giorgio Serino
Abstract:
In this paper, the results of experimental tests performed on a Helical Wire Rope Isolator (HWRI) are presented in order to describe the dynamic and static behavior of the selected metal device in three different displacements ranges, namely small, relatively large, and large displacements ranges, without and under the effect of a vertical load. A testing machine, allowing to apply horizontal displacement or load histories to the tested bearing with a constant vertical load, has been adopted to perform the dynamic and static tests. According to the experimental results, the dynamic behavior of the tested device depends on the applied displacement amplitude. Indeed, the HWRI displays a softening and a hardening stiffness at small and relatively large displacements, respectively, and a stronger nonlinear stiffening behavior at large displacements. Furthermore, the experimental tests reveal that the application of a vertical load allows to have a more flexible device with higher damping properties and that the applied vertical load affects much less the dynamic response of the metal device at large displacements. Finally, a decrease in the static to dynamic effective stiffness ratio with increasing displacement amplitude has been observed.Keywords: base isolation, earthquake engineering, experimental hysteresis loops, wire rope isolators
Procedia PDF Downloads 432868 Performance Evaluation of Iar Multi Crop Thresher
Authors: Idris Idris Sunusi, U.S. Muhammed, N.A. Sale, I.B. Dalha, N.A. Adam
Abstract:
Threshing efficiency and mechanical grain damages are among the important parameters used in rating the performance of agricultural threshers. To be acceptable to farmers, threshers should have high threshing efficiency and low grain. The objective of the research is to evaluate the performances of the thresher using sorghum and millet, the performances parameters considered are; threshing efficiency and mechanical grain damage. For millet, four drum speed levels; 700, 800, 900 and 1000 rpm were considered while for sorghum; 600, 700, 800 and 900 rpm were considered. The feed rate levels were 3, 4, 5 and 6 kg/min for both sorghum and millet; the levels of moisture content were 8.93 and 10.38% for sorghum and 9.21 and 10.81% for millet. For millet the test result showed a maximum of 98.37 threshing efficiencies and a minimum of 0.24% mechanical grain damage while for sorghum the test result indicated a maximum of 99.38 threshing efficiencies, and a minimum of 0.75% mechanical grain damage. In comparison to the previous thresher, the threshing efficiency and mechanical grain damage of the modified machine has improved by 2.01% and 330.56% for millet and 5.31%, 287.64% for sorghum. Also analysis of variance (ANOVA) showed that, the effect of drum speed, feed rate and moisture content were significant on the performance parameters.Keywords: Threshing Efficiency, Mechanical Grain Damages, Sorghum and Millet, Multi Crop Thresher
Procedia PDF Downloads 348867 When the ‘Buddha’s Tree Itself Becomes a Rhizome’: The Religious Itinerant, Nomad Science and the Buddhist State
Authors: James Taylor
Abstract:
This paper considers the political, geo-philosophical musings of Deleuze and Guattari on spatialisation, place and movement in relation to the religious nomad (wandering ascetics and reclusive forest monks) inhabiting the borderlands of Thailand. A nomadic science involves improvised ascetic practices between the molar lines striated by modern state apparatuses. The wandering ascetics, inhabiting a frontier political ecology, stand in contrast to the appropriating, sedentary metaphysics and sanctifying arborescence of statism and its corollary place-making, embedded in rootedness and territorialisation. It is argued that the religious nomads, residing on the endo-exteriorities of the state, came to represent a rhizomatic and politico-ontological threat to centre-nation and its apparatus of capture. The paper also theorises transitions and movement at the borderlands in the context of the state’s monastic reforms. These reforms, and its pervasive royal science, problematised the interstitial zones of the early ascetic wanderers in their radical cross-cutting networks and lines, moving within and across demarcated frontiers. Indeed, the ascetic wanderers and their allegorical war machine were seen as a source of wild, free-floating charisma and mystical power, eventually appropriated by the centre-nation in it’s becoming unitary and fixed.Keywords: Deleuze and Guattari, religious nomad, centre-nation, borderlands, Buddhism
Procedia PDF Downloads 83866 A Framework for Auditing Multilevel Models Using Explainability Methods
Authors: Debarati Bhaumik, Diptish Dey
Abstract:
Multilevel models, increasingly deployed in industries such as insurance, food production, and entertainment within functions such as marketing and supply chain management, need to be transparent and ethical. Applications usually result in binary classification within groups or hierarchies based on a set of input features. Using open-source datasets, we demonstrate that popular explainability methods, such as SHAP and LIME, consistently underperform inaccuracy when interpreting these models. They fail to predict the order of feature importance, the magnitudes, and occasionally even the nature of the feature contribution (negative versus positive contribution to the outcome). Besides accuracy, the computational intractability of SHAP for binomial classification is a cause of concern. For transparent and ethical applications of these hierarchical statistical models, sound audit frameworks need to be developed. In this paper, we propose an audit framework for technical assessment of multilevel regression models focusing on three aspects: (i) model assumptions & statistical properties, (ii) model transparency using different explainability methods, and (iii) discrimination assessment. To this end, we undertake a quantitative approach and compare intrinsic model methods with SHAP and LIME. The framework comprises a shortlist of KPIs, such as PoCE (Percentage of Correct Explanations) and MDG (Mean Discriminatory Gap) per feature, for each of these three aspects. A traffic light risk assessment method is furthermore coupled to these KPIs. The audit framework will assist regulatory bodies in performing conformity assessments of AI systems using multilevel binomial classification models at businesses. It will also benefit businesses deploying multilevel models to be future-proof and aligned with the European Commission’s proposed Regulation on Artificial Intelligence.Keywords: audit, multilevel model, model transparency, model explainability, discrimination, ethics
Procedia PDF Downloads 90865 Cognitive Science Based Scheduling in Grid Environment
Authors: N. D. Iswarya, M. A. Maluk Mohamed, N. Vijaya
Abstract:
Grid is infrastructure that allows the deployment of distributed data in large size from multiple locations to reach a common goal. Scheduling data intensive applications becomes challenging as the size of data sets are very huge in size. Only two solutions exist in order to tackle this challenging issue. First, computation which requires huge data sets to be processed can be transferred to the data site. Second, the required data sets can be transferred to the computation site. In the former scenario, the computation cannot be transferred since the servers are storage/data servers with little or no computational capability. Hence, the second scenario can be considered for further exploration. During scheduling, transferring huge data sets from one site to another site requires more network bandwidth. In order to mitigate this issue, this work focuses on incorporating cognitive science in scheduling. Cognitive Science is the study of human brain and its related activities. Current researches are mainly focused on to incorporate cognitive science in various computational modeling techniques. In this work, the problem solving approach of human brain is studied and incorporated during the data intensive scheduling in grid environments. Here, a cognitive engine is designed and deployed in various grid sites. The intelligent agents present in CE will help in analyzing the request and creating the knowledge base. Depending upon the link capacity, decision will be taken whether to transfer data sets or to partition the data sets. Prediction of next request is made by the agents to serve the requesting site with data sets in advance. This will reduce the data availability time and data transfer time. Replica catalog and Meta data catalog created by the agents assist in decision making process.Keywords: data grid, grid workflow scheduling, cognitive artificial intelligence
Procedia PDF Downloads 392864 Traffic Forecasting for Open Radio Access Networks Virtualized Network Functions in 5G Networks
Authors: Khalid Ali, Manar Jammal
Abstract:
In order to meet the stringent latency and reliability requirements of the upcoming 5G networks, Open Radio Access Networks (O-RAN) have been proposed. The virtualization of O-RAN has allowed it to be treated as a Network Function Virtualization (NFV) architecture, while its components are considered Virtualized Network Functions (VNFs). Hence, intelligent Machine Learning (ML) based solutions can be utilized to apply different resource management and allocation techniques on O-RAN. However, intelligently allocating resources for O-RAN VNFs can prove challenging due to the dynamicity of traffic in mobile networks. Network providers need to dynamically scale the allocated resources in response to the incoming traffic. Elastically allocating resources can provide a higher level of flexibility in the network in addition to reducing the OPerational EXpenditure (OPEX) and increasing the resources utilization. Most of the existing elastic solutions are reactive in nature, despite the fact that proactive approaches are more agile since they scale instances ahead of time by predicting the incoming traffic. In this work, we propose and evaluate traffic forecasting models based on the ML algorithm. The algorithms aim at predicting future O-RAN traffic by using previous traffic data. Detailed analysis of the traffic data was carried out to validate the quality and applicability of the traffic dataset. Hence, two ML models were proposed and evaluated based on their prediction capabilities.Keywords: O-RAN, traffic forecasting, NFV, ARIMA, LSTM, elasticity
Procedia PDF Downloads 221863 Teacher’s Personality Potential Contributes to Personality Development and Well-being of Schoolchildren: A Longitudinal Study in Russia
Authors: Elena G. Diryugina, Maria A. Dovger, Maria V. Lunkina, Alexandra A. Ianchenko
Abstract:
The personality development and well-being of children have become important focuses of school education and indicators of its quality. The studies show that academic success depends more on personality and motivation than on intelligence and giftedness. Those personality resources that help a person to maintain well-being both here and now and in the future constitute their personality potential. The development of schoolchildrens' personality potential can help them meet the challenges of the modern world and achieve new educational goals. At the same time, it is noted that the pedagogical factor is one of the most significant in relation to schoolchildrens' success and well-being. What is important for teachers to develop in order to make their students feel more competent and maintain well-being? As part of the Developmental Environment Programme of the Charitable Foundation ‘Investment in the Future’, a longitudinal study of the personality potential and well-being of educators and schoolchildren was conducted from 2018 to 2023. More than 2,500 teachers and over 4,000 students from Russia took part. It was found that behind a teacher's communication style, an important construct that influences the motivation of schoolchildren and the satisfaction of their basic psychological needs, is the personal potential of that teacher. Their personality potential correlates with the social-emotional development of schoolchildren in junior grades. A teacher's communication style with adolescents contributes to their academic motivation, self-esteem and satisfaction with life and learning. In addition, child well-being cannot be promoted in isolation from attention to the psychological well-being of teachers. Their social well-being and engagement are higher when they are included in professional learning communities. The results will be helpful for both positive education researchers and practitioners to identify an approach to child personality development and well-being that is achieved primarily through the personality development and well-being of school staff members and mostly teachers.Keywords: Personality development, personality potential, schoolchildren, teaching style, well-being
Procedia PDF Downloads 43862 Development and Power Characterization of an IoT Network for Agricultural Imaging Applications
Authors: Jacob Wahl, Jane Zhang
Abstract:
This paper describes the development and characterization of a prototype IoT network for use with agricultural imaging and monitoring applications. The sensor and gateway nodes are designed using the ESP32 SoC with integrated Bluetooth Low Energy 4.2 and Wi-Fi. A development board, the Arducam IoTai ESP32, is used for prototyping, testing, and power measurements. Google’s Firebase is used as the cloud storage site for image data collected by the sensor. The sensor node captures images using the OV2640 2MP camera module and transmits the image data to the gateway via Bluetooth Low Energy. The gateway then uploads the collected images to Firebase via a known nearby Wi-Fi network connection. This image data can then be processed and analyzed by computer vision and machine learning pipelines to assess crop growth or other needs. The sensor node achieves a wireless transmission data throughput of 220kbps while consuming 150mA of current; the sensor sleeps at 162µA. The sensor node device lifetime is estimated to be 682 days on a 6600mAh LiPo battery while acquiring five images per day based on the development board power measurements. This network can be utilized by any application that requires high data rates, low power consumption, short-range communication, and large amounts of data to be transmitted at low-frequency intervals.Keywords: Bluetooth low energy, ESP32, firebase cloud, IoT, smart farming
Procedia PDF Downloads 137861 The Analysis Fleet Operational Performance as an Indicator of Load and Haul Productivity
Authors: Linet Melisa Daubanes, Nhleko Monique Chiloane
Abstract:
The shovel-truck system is the most prevalent material handling system used in surface mining operations. Material handling entails the loading and hauling of material from production areas to dumping areas. The material handling process has operational delays that have a negative impact on the productivity of the load and haul fleet. Factors that may contribute to operational delays include shovel-truck mismatch, haul routes, machine breakdowns, extreme weather conditions, etc. The aim of this paper is to investigate factors that contribute to operational delays affecting the productivity of the load and haul fleet at the mine. Productivity is the measure of the effectiveness of producing products from a given quantity of units, the ratio of output to inputs. Productivity can be improved by producing more outputs with the same or fewer units and/or introducing better working methods etc. Several key performance indicators (KPI) for the evaluation of productivity will be discussed in this study. These KPIs include but are not limited to hauling conditions, bucket fill factor, cycle time, and utilization. The research methodology of this study is a combination of on-site time studies and observations. Productivity can be optimized by managing the factors that affect the operational performance of the haulage fleet.Keywords: cycle time, fleet performance, load and haul, surface mining
Procedia PDF Downloads 194860 Random Forest Classification for Population Segmentation
Authors: Regina Chua
Abstract:
To reduce the costs of re-fielding a large survey, a Random Forest classifier was applied to measure the accuracy of classifying individuals into their assigned segments with the fewest possible questions. Given a long survey, one needed to determine the most predictive ten or fewer questions that would accurately assign new individuals to custom segments. Furthermore, the solution needed to be quick in its classification and usable in non-Python environments. In this paper, a supervised Random Forest classifier was modeled on a dataset with 7,000 individuals, 60 questions, and 254 features. The Random Forest consisted of an iterative collection of individual decision trees that result in a predicted segment with robust precision and recall scores compared to a single tree. A random 70-30 stratified sampling for training the algorithm was used, and accuracy trade-offs at different depths for each segment were identified. Ultimately, the Random Forest classifier performed at 87% accuracy at a depth of 10 with 20 instead of 254 features and 10 instead of 60 questions. With an acceptable accuracy in prioritizing feature selection, new tools were developed for non-Python environments: a worksheet with a formulaic version of the algorithm and an embedded function to predict the segment of an individual in real-time. Random Forest was determined to be an optimal classification model by its feature selection, performance, processing speed, and flexible application in other environments.Keywords: machine learning, supervised learning, data science, random forest, classification, prediction, predictive modeling
Procedia PDF Downloads 91