Search results for: feature detection and description
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5433

Search results for: feature detection and description

2283 Engineering Thermal-Hydraulic Simulator Based on Complex Simulation Suite “Virtual Unit of Nuclear Power Plant”

Authors: Evgeny Obraztsov, Ilya Kremnev, Vitaly Sokolov, Maksim Gavrilov, Evgeny Tretyakov, Vladimir Kukhtevich, Vladimir Bezlepkin

Abstract:

Over the last decade, a specific set of connected software tools and calculation codes has been gradually developed. It allows simulating I&C systems, thermal-hydraulic, neutron-physical and electrical processes in elements and systems at the Unit of NPP (initially with WWER (pressurized water reactor)). In 2012 it was called a complex simulation suite “Virtual Unit of NPP” (or CSS “VEB” for short). Proper application of this complex tool should result in a complex coupled mathematical computational model. And for a specific design of NPP, it is called the Virtual Power Unit (or VPU for short). VPU can be used for comprehensive modelling of a power unit operation, checking operator's functions on a virtual main control room, and modelling complicated scenarios for normal modes and accidents. In addition, CSS “VEB” contains a combination of thermal hydraulic codes: the best-estimate (two-liquid) calculation codes KORSAR and CORTES and a homogenous calculation code TPP. So to analyze a specific technological system one can build thermal-hydraulic simulation models with different detalization levels up to a nodalization scheme with real geometry. And the result at some points is similar to the notion “engineering/testing simulator” described by the European utility requirements (EUR) for LWR nuclear power plants. The paper is dedicated to description of the tools mentioned above and an example of the application of the engineering thermal-hydraulic simulator in analysis of the boron acid concentration in the primary coolant (changed by the make-up and boron control system).

Keywords: best-estimate code, complex simulation suite, engineering simulator, power plant, thermal hydraulic, VEB, virtual power unit

Procedia PDF Downloads 368
2282 Design and Development of 5-DOF Color Sorting Manipulator for Industrial Applications

Authors: Atef A. Ata, Sohair F. Rezeka, Ahmed El-Shenawy, Mohammed Diab

Abstract:

Image processing in today’s world grabs massive attentions as it leads to possibilities of broaden application in many fields of high technology. The real challenge is how to improve existing sorting system applications which consists of two integrated stations of processing and handling with a new image processing feature. Existing color sorting techniques use a set of inductive, capacitive, and optical sensors to differentiate object color. This research presents a mechatronics color sorting system solution with the application of image processing. A 5-DOF robot arm is designed and developed with pick and place operation to be main part of the color sorting system. Image processing procedure senses the circular objects in an image captured in real time by a webcam attached at the end-effector then extracts color and position information out of it. This information is passed as a sequence of sorting commands to the manipulator that has pick-and-place mechanism. Performance analysis proves that this color based object sorting system works very accurate under ideal condition in term of adequate illumination, circular objects shape and color. The circular objects tested for sorting are red, green and blue. For non-ideal condition, such as unspecified color the accuracy reduces to 80%.

Keywords: robotics manipulator, 5-DOF manipulator, image processing, color sorting, pick-and-place

Procedia PDF Downloads 360
2281 Establishment of a Classifier Model for Early Prediction of Acute Delirium in Adult Intensive Care Unit Using Machine Learning

Authors: Pei Yi Lin

Abstract:

Objective: The objective of this study is to use machine learning methods to build an early prediction classifier model for acute delirium to improve the quality of medical care for intensive care patients. Background: Delirium is a common acute and sudden disturbance of consciousness in critically ill patients. After the occurrence, it is easy to prolong the length of hospital stay and increase medical costs and mortality. In 2021, the incidence of delirium in the intensive care unit of internal medicine was as high as 59.78%, which indirectly prolonged the average length of hospital stay by 8.28 days, and the mortality rate is about 2.22% in the past three years. Therefore, it is expected to build a delirium prediction classifier through big data analysis and machine learning methods to detect delirium early. Method: This study is a retrospective study, using the artificial intelligence big data database to extract the characteristic factors related to delirium in intensive care unit patients and let the machine learn. The study included patients aged over 20 years old who were admitted to the intensive care unit between May 1, 2022, and December 31, 2022, excluding GCS assessment <4 points, admission to ICU for less than 24 hours, and CAM-ICU evaluation. The CAMICU delirium assessment results every 8 hours within 30 days of hospitalization are regarded as an event, and the cumulative data from ICU admission to the prediction time point are extracted to predict the possibility of delirium occurring in the next 8 hours, and collect a total of 63,754 research case data, extract 12 feature selections to train the model, including age, sex, average ICU stay hours, visual and auditory abnormalities, RASS assessment score, APACHE-II Score score, number of invasive catheters indwelling, restraint and sedative and hypnotic drugs. Through feature data cleaning, processing and KNN interpolation method supplementation, a total of 54595 research case events were extracted to provide machine learning model analysis, using the research events from May 01 to November 30, 2022, as the model training data, 80% of which is the training set for model training, and 20% for the internal verification of the verification set, and then from December 01 to December 2022 The CU research event on the 31st is an external verification set data, and finally the model inference and performance evaluation are performed, and then the model has trained again by adjusting the model parameters. Results: In this study, XG Boost, Random Forest, Logistic Regression, and Decision Tree were used to analyze and compare four machine learning models. The average accuracy rate of internal verification was highest in Random Forest (AUC=0.86), and the average accuracy rate of external verification was in Random Forest and XG Boost was the highest, AUC was 0.86, and the average accuracy of cross-validation was the highest in Random Forest (ACC=0.77). Conclusion: Clinically, medical staff usually conduct CAM-ICU assessments at the bedside of critically ill patients in clinical practice, but there is a lack of machine learning classification methods to assist ICU patients in real-time assessment, resulting in the inability to provide more objective and continuous monitoring data to assist Clinical staff can more accurately identify and predict the occurrence of delirium in patients. It is hoped that the development and construction of predictive models through machine learning can predict delirium early and immediately, make clinical decisions at the best time, and cooperate with PADIS delirium care measures to provide individualized non-drug interventional care measures to maintain patient safety, and then Improve the quality of care.

Keywords: critically ill patients, machine learning methods, delirium prediction, classifier model

Procedia PDF Downloads 58
2280 Automatic Classification for the Degree of Disc Narrowing from X-Ray Images Using CNN

Authors: Kwangmin Joo

Abstract:

Automatic detection of lumbar vertebrae and classification method is proposed for evaluating the degree of disc narrowing. Prior to classification, deep learning based segmentation is applied to detect individual lumbar vertebra. M-net is applied to segment five lumbar vertebrae and fine-tuning segmentation is employed to improve the accuracy of segmentation. Using the features extracted from previous step, clustering technique, k-means clustering, is applied to estimate the degree of disc space narrowing under four grade scoring system. As preliminary study, techniques proposed in this research could help building an automatic scoring system to diagnose the severity of disc narrowing from X-ray images.

Keywords: Disc space narrowing, Degenerative disc disorders, Deep learning based segmentation, Clustering technique

Procedia PDF Downloads 116
2279 Machine Learning for Aiding Meningitis Diagnosis in Pediatric Patients

Authors: Karina Zaccari, Ernesto Cordeiro Marujo

Abstract:

This paper presents a Machine Learning (ML) approach to support Meningitis diagnosis in patients at a children’s hospital in Sao Paulo, Brazil. The aim is to use ML techniques to reduce the use of invasive procedures, such as cerebrospinal fluid (CSF) collection, as much as possible. In this study, we focus on predicting the probability of Meningitis given the results of a blood and urine laboratory tests, together with the analysis of pain or other complaints from the patient. We tested a number of different ML algorithms, including: Adaptative Boosting (AdaBoost), Decision Tree, Gradient Boosting, K-Nearest Neighbors (KNN), Logistic Regression, Random Forest and Support Vector Machines (SVM). Decision Tree algorithm performed best, with 94.56% and 96.18% accuracy for training and testing data, respectively. These results represent a significant aid to doctors in diagnosing Meningitis as early as possible and in preventing expensive and painful procedures on some children.

Keywords: machine learning, medical diagnosis, meningitis detection, pediatric research

Procedia PDF Downloads 140
2278 Design and Control of a Knee Rehabilitation Device Using an MR-Fluid Brake

Authors: Mina Beheshti, Vida Shams, Mojtaba Esfandiari, Farzaneh Abdollahi, Abdolreza Ohadi

Abstract:

Most of the people who survive a stroke need rehabilitation tools to regain their mobility. The core function of these devices is a brake actuator. The goal of this study is to design and control a magnetorheological brake which can be used as a rehabilitation tool. In fact, the fluid used in this brake is called magnetorheological fluid or MR that properties can change by variation of the magnetic field. The braking properties can be set as control by using this feature of the fluid. In this research, different MR brake designs are first introduced in each design, and the dimensions of the brake have been determined based on the required torque for foot movement. To calculate the brake dimensions, it is assumed that the shear stress distribution in the fluid is uniform and the fluid is in its saturated state. After designing the rehabilitation brake, the mathematical model of the healthy movement of a healthy person is extracted. Due to the nonlinear nature of the system and its variability, various adaptive controllers, neural networks, and robust have been implemented to estimate the parameters and control the system. After calculating torque and control current, the best type of controller in terms of error and control current has been selected. Finally, this controller is implemented on the experimental data of the patient's movements, and the control current is calculated to achieve the desired torque and motion.

Keywords: rehabilitation, magnetorheological fluid, knee, brake, adaptive control, robust control, neural network control, torque control

Procedia PDF Downloads 141
2277 Mental Accounting Theory Development Review and Application

Authors: Kang-Hsien Li

Abstract:

Along with global industries in using technology to enhance the application, make the study drawn more close to the people’s behavior and produce data analysis, extended out from the mental accounting of prospect theory, this paper provides the marketing and financial applications in the field of exploration and discussions with the future. For the foreseeable future, the payment behavior depends on the form of currency, which affects a variety of product types on the marketing of marketing strategy to provide diverse payment methods to enhance the overall sales performance. This not only affects people's consumption also affects people's investments. Credit card, PayPal, Apple pay, Bitcoin and any other with advances in technology and other emerging payment instruments, began to affect people for the value and the concept of money. Such as the planning of national social welfare policies, monetary and financial regulators and regulators. The expansion can be expected to discuss marketing and finance-related mental problems at the same time, recent studies reflect two different ideas, the first idea is that individuals affected by situational frames, not broad impact at the event level, affected by the people basically mental, second idea is that when an individual event affects a broader range, and majority of people will choose the same at the time that the rational choice. That are applied to practical application of marketing, at the same time provide an explanation in the financial market under the anomalies, due to the financial markets has varied investment products and different market participants, that also highlights these two points. It would provide in-depth description of humanity's mental. Certainly, about discuss mental accounting aspects, while artificial intelligence application development, although people would be able to reduce prejudice decisions, that will also lead to more discussion on the economic and marketing strategy.

Keywords: mental accounting, behavior economics, consumer behaviors, decision-making

Procedia PDF Downloads 445
2276 Mercury Detection in Two Fishes from the Persian Gulf

Authors: Zahra Khoshnood, Mehdi Kazaie, Sajedeh Neisi

Abstract:

In 2013, 24 fish samples were taken from two fishery regions in the north of Persian Gulf near the Iranian coastal lines. The two flatfishes were Yellofin seabream (Acanthopagrus latus) and Longtail tuna (Thannus tonggol). We analyzed total Hg concentration of liver and muscle tissues by Mercury Analyzer (model LECO AMA 254). The average concentration of total Hg in edible Muscle tissue of deep-Flounder was measured in Bandar-Abbas and was found to be 18.92 and it was 10.19 µg.g-1 in Bandar-Lengeh. The corresponding values for Oriental sole were 8.47 and 0.08 µg.g-1. The average concentration of Hg in liver tissue of deep-Flounder, in Bandar-Abbas was 25.49 and that in Bandar-Lengeh was 12.52 µg.g-1.the values for Oriental sole were 11.88 and 3.2 µg.g-1 in Bandar-Abbas and Bandar-Lengeh, respectively.

Keywords: mercury, Acanthopagrus latus, Thannus tonggol, Persian Gulf

Procedia PDF Downloads 591
2275 Chemometric QSRR Evaluation of Behavior of s-Triazine Pesticides in Liquid Chromatography

Authors: Lidija R. Jevrić, Sanja O. Podunavac-Kuzmanović, Strahinja Z. Kovačević

Abstract:

This study considers the selection of the most suitable in silico molecular descriptors that could be used for s-triazine pesticides characterization. Suitable descriptors among topological, geometrical and physicochemical are used for quantitative structure-retention relationships (QSRR) model establishment. Established models were obtained using linear regression (LR) and multiple linear regression (MLR) analysis. In this paper, MLR models were established avoiding multicollinearity among the selected molecular descriptors. Statistical quality of established models was evaluated by standard and cross-validation statistical parameters. For detection of similarity or dissimilarity among investigated s-triazine pesticides and their classification, principal component analysis (PCA) and hierarchical cluster analysis (HCA) were used and gave similar grouping. This study is financially supported by COST action TD1305.

Keywords: chemometrics, classification analysis, molecular descriptors, pesticides, regression analysis

Procedia PDF Downloads 383
2274 Novel Synthesis of Metal Oxide Nanoparticles from Type IV Deep Eutectic Solvents

Authors: Lorenzo Gontrani, Marilena Carbone, Domenica Tommasa Donia, Elvira Maria Bauer, Pietro Tagliatesta

Abstract:

One of the fields where DES shows remarkable added values is the synthesis Of inorganic materials, in particular nanoparticles. In this field, the higher- ent and highly-tunable nano-homogeneities of DES structure give origin to a marked templating effect, a precious role that has led to the recent bloom of a vast number of studies exploiting these new synthesis media to prepare Nanomaterials and composite structures of various kinds. In this contribution, the most recent developments in the field will be reviewed, and some ex-citing examples of novel metal oxide nanoparticles syntheses using non-toxic type-IV Deep Eutectic Solvents will be described. The prepared materials possess nanometric dimensions and show flower-like shapes. The use of the pre- pared nanoparticles as fluorescent materials for the detection of various contaminants is under development.

Keywords: metal deep eutectic solvents, nanoparticles, inorganic synthesis, type IV DES, lamellar

Procedia PDF Downloads 125
2273 Agent-Based Modelling to Improve Dairy-origin Beef Production: Model Description and Evaluation

Authors: Addisu H. Addis, Hugh T. Blair, Paul R. Kenyon, Stephen T. Morris, Nicola M. Schreurs, Dorian J. Garrick

Abstract:

Agent-based modeling (ABM) enables an in silico representation of complex systems and cap-tures agent behavior resulting from interaction with other agents and their environment. This study developed an ABM to represent a pasture-based beef cattle finishing systems in New Zea-land (NZ) using attributes of the rearer, finisher, and processor, as well as specific attributes of dairy-origin beef cattle. The model was parameterized using values representing 1% of NZ dairy-origin cattle, and 10% of rearers and finishers in NZ. The cattle agent consisted of 32% Holstein-Friesian, 50% Holstein-Friesian–Jersey crossbred, and 8% Jersey, with the remainder being other breeds. Rearers and finishers repetitively and simultaneously interacted to determine the type and number of cattle populating the finishing system. Rearers brought in four-day-old spring-born calves and reared them until 60 calves (representing a full truck load) on average had a live weight of 100 kg before selling them on to finishers. Finishers mainly attained weaners from rearers, or directly from dairy farmers when weaner demand was higher than the supply from rearers. Fast-growing cattle were sent for slaughter before the second winter, and the re-mainder were sent before their third winter. The model finished a higher number of bulls than heifers and steers, although it was 4% lower than the industry reported value. Holstein-Friesian and Holstein-Friesian–Jersey-crossbred cattle dominated the dairy-origin beef finishing system. Jersey cattle account for less than 5% of total processed beef cattle. Further studies to include re-tailer and consumer perspectives and other decision alternatives for finishing farms would im-prove the applicability of the model for decision-making processes.

Keywords: agent-based modelling, dairy cattle, beef finishing, rearers, finishers

Procedia PDF Downloads 85
2272 General Mathematical Framework for Analysis of Cattle Farm System

Authors: Krzysztof Pomorski

Abstract:

In the given work we present universal mathematical framework for modeling of cattle farm system that can set and validate various hypothesis that can be tested against experimental data. The presented work is preliminary but it is expected to be valid tool for future deeper analysis that can result in new class of prediction methods allowing early detection of cow dieseaes as well as cow performance. Therefore the presented work shall have its meaning in agriculture models and in machine learning as well. It also opens the possibilities for incorporation of certain class of biological models necessary in modeling of cow behavior and farm performance that might include the impact of environment on the farm system. Particular attention is paid to the model of coupled oscillators that it the basic building hypothesis that can construct the model showing certain periodic or quasiperiodic behavior.

Keywords: coupled ordinary differential equations, cattle farm system, numerical methods, stochastic differential equations

Procedia PDF Downloads 135
2271 Population Stereotype Production, User Factors, and Icon Design for Underserved Communities of Rural India

Authors: Avijit Sengupta, Klarissa Ting Ting Cheng, Maffee Peng-Hui Wan

Abstract:

This study investigates the influence of user factors and referent characteristics on representation types generated using the stereotype production method for designing icons. Sixty-eight participants of farming communities were asked to draw images based on sixteen feature referents. Significant statistical differences were found between the types of representations generated for contextual and context-independent referents. Strong correlations were observed between years of formal education and total number of abstract representations produced for both contextual and context-independent referents. However, representation characteristics were not influenced by other user factors such as participants’ experience with mobile phone and years of farming experience. A statistically significant tendency of making concrete representations was observed for both contextual and context-independent referents. These findings provide insights on community members’ involvement in icon design and suggest a consolidated icon design strategy based on population stereotype, particularly for under-served rural communities of India.

Keywords: abstract representation, concrete representation, participatory design, population stereotype

Procedia PDF Downloads 369
2270 Insufficiency Fracture of Femoral Head in Patients Treated With Intramedullary Nailing for Proximal Femur Fracture

Authors: Jai Hyung Park, Eugene Kim, Jin Hun Park, Min Joon Oh

Abstract:

Introduction: Subchondral insufficiency fracture of the femoral head (SIF) is a rare complication; however, it has been recognized to cause femoral head collapse. Subchondral insufficiency fracture (SIF) is caused by normal or physiological stress without any trauma. It has been reported in osteoporotic patients after the fixation of the proximal femur with an Intramedullary nail. Case presentation: We reported 5 cases with SIF of the femoral head after proximal femur fracture fixation with Intra-medullary nail. All patients had osteoporosis as an underlying disease. Good reduction was achieved in all 5 patients. SIF was found from about 3 months to 4 years after the initial operation, and all the fractures were solidly united at the final diagnosis. We investigated retrospectively the feature of those cases and several factors that affected the occurrence of SIF. Discussion: There are a few discussions regarding the SIF of the femoral head. These discussions may include the predisposing risk factors, how to diagnose the SIF in osteoporotic patients, and the peri-operative factors to prevent SIF. Conclusion: Subchondral insufficiency fracture of the femoral head is a considerable complication after the internal fixation of the proximal femur. There are several factors that can be modified. If they could be controlled in the peri-operative period, SIF could be prevented or handled in advance. Other options related to arthroplasty can be considered in old osteoporotic patients.

Keywords: insufficiency fracture of femoral head, intra-medullary nail, osteoporosis, proximal femur fracture

Procedia PDF Downloads 112
2269 An Algorithm for Removal of Noise from X-Ray Images

Authors: Sajidullah Khan, Najeeb Ullah, Wang Yin Chai, Chai Soo See

Abstract:

In this paper, we propose an approach to remove impulse and Poisson noise from X-ray images. Many filters have been used for impulse noise removal from color and gray scale images with their own strengths and weaknesses but X-ray images contain Poisson noise and unfortunately there is no intelligent filter which can detect impulse and Poisson noise from X-ray images. Our proposed filter uses the upgraded layer discrimination approach to detect both Impulse and Poisson noise corrupted pixels in X-ray images and then restores only those detected pixels with a simple efficient and reliable one line equation. Our Proposed algorithms are very effective and much more efficient than all existing filters used only for Impulse noise removal. The proposed method uses a new powerful and efficient noise detection method to determine whether the pixel under observation is corrupted or noise free. Results from computer simulations are used to demonstrate pleasing performance of our proposed method.

Keywords: X-ray image de-noising, impulse noise, poisson noise, PRWF

Procedia PDF Downloads 370
2268 Remote Patient Monitoring for Covid-19

Authors: Launcelot McGrath

Abstract:

The Coronavirus disease 2019 (COVID-19) has spread rapidly around the world, resulting in high mortality rates and very large numbers of people requiring medical treatment in ICU. Management of patient hospitalisation is a critical aspect to control this disease and reduce chaos in the healthcare systems. Remote monitoring provides a solution to protect vulnerable and elderly high-risk patients. Continuous remote monitoring of oxygen saturation, respiratory rate, heart rate, and temperature, etc., provides medical systems with up-to-the-minute information about their patients' statuses. Remote monitoring also limits the spread of infection by reducing hospital overcrowding. This paper examines the potential of remote monitoring for Covid-19 to assist in the rapid identification of patients at risk, facilitate the detection of patient deterioration, and enable early interventions.

Keywords: remote monitoring, patient care, oxygen saturation, Covid-19, hospital management

Procedia PDF Downloads 94
2267 Fully Eulerian Finite Element Methodology for the Numerical Modeling of the Dynamics of Heart Valves

Authors: Aymen Laadhari

Abstract:

During the last decade, an increasing number of contributions have been made in the fields of scientific computing and numerical methodologies applied to the study of the hemodynamics in the heart. In contrast, the numerical aspects concerning the interaction of pulsatile blood flow with highly deformable thin leaflets have been much less explored. This coupled problem remains extremely challenging and numerical difficulties include e.g. the resolution of full Fluid-Structure Interaction problem with large deformations of extremely thin leaflets, substantial mesh deformations, high transvalvular pressure discontinuities, contact between leaflets. Although the Lagrangian description of the structural motion and strain measures is naturally used, many numerical complexities can arise when studying large deformations of thin structures. Eulerian approaches represent a promising alternative to readily model large deformations and handle contact issues. We present a fully Eulerian finite element methodology tailored for the simulation of pulsatile blood flow in the aorta and sinus of Valsalva interacting with highly deformable thin leaflets. Our method enables to use a fluid solver on a fixed mesh, whilst being able to easily model the mechanical properties of the valve. We introduce a semi-implicit time integration scheme based on a consistent NewtonRaphson linearization. A variant of the classical Newton method is introduced and guarantees a third-order convergence. High-fidelity computational geometries are built and simulations are performed under physiological conditions. We address in detail the main features of the proposed method, and we report several experiments with the aim of illustrating its accuracy and efficiency.

Keywords: eulerian, level set, newton, valve

Procedia PDF Downloads 272
2266 Friend or Foe: Decoding the Legal Challenges Posed by Artificial Intellegence in the Era of Intellectual Property

Authors: Latika Choudhary

Abstract:

“The potential benefits of Artificial Intelligence are huge, So are the dangers.” - Dave Water. Artificial intelligence is one of the facet of Information technology domain which despite several attempts does not have a clear definition or ambit. However it can be understood as technology to solve problems via automated decisions and predictions. Artificial intelligence is essentially an algorithm based technology which analyses the large amounts of data and then solves problems by detecting useful patterns. Owing to its automated feature it will not be wrong to say that humans & AI have more utility than humans alone or computers alone.1 For many decades AI experienced enthusiasm as well as setbacks, yet it has today become part and parcel of our everyday life, making it convenient or at times problematic. AI and related technology encompass Intellectual Property in multiple ways, the most important being AI technology for management of Intellectual Property, IP for protecting AI and IP as a hindrance to the transparency of AI systems. Thus the relationship between the two is of reciprocity as IP influences AI and vice versa. While AI is a recent concept, the IP laws for protection or even dealing with its challenges are relatively older, raising the need for revision to keep up with the pace of technological advancements. This paper will analyze the relationship between AI and IP to determine how beneficial or conflictual the same is, address how the old concepts of IP are being stretched to its maximum limits so as to accommodate the unwanted consequences of the Artificial Intelligence and propose ways to mitigate the situation so that AI becomes the friend it is and not turn into a potential foe it appears to be.

Keywords: intellectual property rights, information technology, algorithm, artificial intelligence

Procedia PDF Downloads 79
2265 Compilation of Islamic Law as Law Applied Religious Courts in Indonesia (Responding to Changes in Religious Courts Authority)

Authors: Hamdan Arief Hanif, Rahmat Sidiq

Abstract:

Indonesia is a country of law, the legal system adopted by Indonesia is a civil law system. A major feature of the civil law is the codified legislation. Meanwhile the majority of society Indonesia are Muslims, whilst Islamic law itself having the sources written in Qur'an, Sunnah and the opinion of Muslim scholars, generally not codified in book form of legislation that is easy on the set as a reference. in Indonesia, many scholars have different opinions in decisions so that there is no legal certainty in Muslim civil cases, so the need for legal codification, which, as the source of the judges in deciding a case, especially a case in religious courts. This paper raised the topic of discussion which offers a solution to the application of the codification of the Islamic Law which became the core resources in delivering a verdict against Islamic civil related issue; codification usually called a compilation of Islamic Law. Compilation of Islamic Law is highly recommended as a core reference for the judges in religious courts in Indonesia. This compilation which includes a collection of large number of opinions scholars (book of fiqh) that existed previously and are ripened in deduce in order to unify the existing differences. This paper also discusses how the early formation of the compilation and as the right solution in order to create legal certainty and justice especially for the muslim community in Indonesia.

Keywords: Islamic law, compilation, law applied core, religious court

Procedia PDF Downloads 351
2264 Easy Way of Optimal Process-Storage Network Design

Authors: Gyeongbeom Yi

Abstract:

The purpose of this study is to introduce the analytic solution for determining the optimal capacity (lot-size) of a multiproduct, multistage production and inventory system to meet the finished product demand. Reasonable decision-making about the capacity of processes and storage units is an important subject for industry. The industrial solution for this subject is to use the classical economic lot sizing method, EOQ/EPQ (Economic Order Quantity/Economic Production Quantity) model, incorporated with practical experience. However, the unrealistic material flow assumption of the EOQ/EPQ model is not suitable for chemical plant design with highly interlinked processes and storage units. This study overcomes the limitation of the classical lot sizing method developed on the basis of the single product and single stage assumption. The superstructure of the plant considered consists of a network of serially and/or parallelly interlinked processes and storage units. The processes involve chemical reactions with multiple feedstock materials and multiple products as well as mixing, splitting or transportation of materials. The objective function for optimization is minimizing the total cost composed of setup and inventory holding costs as well as the capital costs of constructing processes and storage units. A novel production and inventory analysis method, PSW (Periodic Square Wave) model, is applied. The advantage of the PSW model comes from the fact that the model provides a set of simple analytic solutions in spite of a realistic description of the material flow between processes and storage units. The resulting simple analytic solution can greatly enhance the proper and quick investment decision for plant design and operation problem confronted in diverse economic situations.

Keywords: analytic solution, optimal design, process-storage network

Procedia PDF Downloads 323
2263 A Comparative Assessment of Industrial Composites Using Thermography and Ultrasound

Authors: Mosab Alrashed, Wei Xu, Stephen Abineri, Yifan Zhao, Jörn Mehnen

Abstract:

Thermographic inspection is a relatively new technique for Non-Destructive Testing (NDT) which has been gathering increasing interest due to its relatively low cost hardware and extremely fast data acquisition properties. This technique is especially promising in the area of rapid automated damage detection and quantification. In collaboration with a major industry partner from the aerospace sector advanced thermography-based NDT software for impact damaged composites is introduced. The software is based on correlation analysis of time-temperature profiles in combination with an image enhancement process. The prototype software is aiming to a) better visualise the damages in a relatively easy-to-use way and b) automatically and quantitatively measure the properties of the degradation. Knowing that degradation properties play an important role in the identification of degradation types, tests and results on specimens which were artificially damaged have been performed and analyzed.

Keywords: NDT, correlation analysis, image processing, damage, inspection

Procedia PDF Downloads 532
2262 A Secure Routing Algorithm for ‎Underwater Wireless Sensor Networks

Authors: Seyed Mahdi Jameii

Abstract:

Underwater wireless sensor networks have been attracting the interest of many ‎researchers lately, and the past three decades have beheld the rapid progress of ‎underwater acoustic communication. One of the major problems in underwater wireless ‎sensor networks is how to transfer data from the moving node to the base stations and ‎choose the optimized route for data transmission. Secure routing in underwater ‎wireless sensor network (UWCNs) is necessary for packet delivery. Some routing ‎protocols are proposed for underwater wireless sensor networks. However, a few ‎researches have been done on secure routing in underwater sensor networks. In this ‎article, a secure routing protocol is provided to resist against wormhole and sybil ‎attacks. The results indicated acceptable performance in terms of increasing the packet ‎delivery ratio with regards to the attacks, increasing network lifetime by creating ‎balance in the network energy consumption, high detection rates against the attacks, ‎and low-end to end delay.‎

Keywords: attacks, routing, security, underwater wireless sensor networks

Procedia PDF Downloads 407
2261 Designing an Intelligent Voltage Instability System in Power Distribution Systems in the Philippines Using IEEE 14 Bus Test System

Authors: Pocholo Rodriguez, Anne Bernadine Ocampo, Ian Benedict Chan, Janric Micah Gray

Abstract:

The state of an electric power system may be classified as either stable or unstable. The borderline of stability is at any condition for which a slight change in an unfavourable direction of any pertinent quantity will cause instability. Voltage instability in power distribution systems could lead to voltage collapse and thus power blackouts. The researchers will present an intelligent system using back propagation algorithm that can detect voltage instability and output voltage of a power distribution and classify it as stable or unstable. The researchers’ work is the use of parameters involved in voltage instability as input parameters to the neural network for training and testing purposes that can provide faster detection and monitoring of the power distribution system.

Keywords: back-propagation algorithm, load instability, neural network, power distribution system

Procedia PDF Downloads 426
2260 Diversity and Taxonomy: Malaysian Marine Algae Genus Halimeda (Halimedaceae, Chlorophyta)

Authors: Nur Farah Ain Zainee, Ahmad Ismail, Nazlina Ibrahim, Asmida Ismail

Abstract:

The study of genus Halimeda in Malaysia is in the early stage due to less specific study on its taxonomy. Most of the previous research tend to choose other genus such as Caulerpa and Gracilaria because of the potential of being utilized. The identification of Halimeda is complex by the high morphological variation within individual species due to different types of habitat and the changes in composition of seawater. The study was completed to study the diversity and distribution of Halimeda in Malaysia and to identify the morphological and anatomical differences between Halimeda species. The methods which have been used for this study are collection of Halimeda and seawater, preservation of specimen, identification of the specimen including the preparation of the temporary slide and decalcification of the calcium layer by using diluted hydrochloric acid. The specimen were processed in laboratory and kept as herbarium specimen in Algae Herbarium, Universiti Kebangsaan Malaysia. Environmental parameters were tested by using YSI multiparameter probe and the recorded data were temperature, salinity, pH and dissolved oxygen. The nutrient content of seawater such as nitrate and phosphate were analysed by using Hach kit model DR 2000. In the present study, out of 330 herbarium specimen, ten species were identified as Halimeda cuneata, H. discoidea, H. macroloba, H. macrophysa, H. opuntia, H. simulans, H. stuposa, H. taenicola, H. tuna and H. velasquezii. Of these, five species were new record to Malaysia. They are Halimeda cuneata, H. macrophysa, H. stuposa, H. taenicola and H. velasquezii. H. opuntia was found as the most abundance species with wide distribution in Malaysia coastal area. Meanwhile, from the study of their distribution, two localities in which Pulau Balak Balak, Kudat and Pulau Langkawi, Kedah, were noted having high number of Halimeda species. As a conclusion, this study has successfully identified ten species of Halimeda of Malaysia with full description of morphological characteristics that may assist further researcher to differentiate and identify Halimeda.

Keywords: Distribution, diversity, Halimeda, morphological, taxonomy

Procedia PDF Downloads 330
2259 3D Classification Optimization of Low-Density Airborne Light Detection and Ranging Point Cloud by Parameters Selection

Authors: Baha Eddine Aissou, Aichouche Belhadj Aissa

Abstract:

Light detection and ranging (LiDAR) is an active remote sensing technology used for several applications. Airborne LiDAR is becoming an important technology for the acquisition of a highly accurate dense point cloud. A classification of airborne laser scanning (ALS) point cloud is a very important task that still remains a real challenge for many scientists. Support vector machine (SVM) is one of the most used statistical learning algorithms based on kernels. SVM is a non-parametric method, and it is recommended to be used in cases where the data distribution cannot be well modeled by a standard parametric probability density function. Using a kernel, it performs a robust non-linear classification of samples. Often, the data are rarely linearly separable. SVMs are able to map the data into a higher-dimensional space to become linearly separable, which allows performing all the computations in the original space. This is one of the main reasons that SVMs are well suited for high-dimensional classification problems. Only a few training samples, called support vectors, are required. SVM has also shown its potential to cope with uncertainty in data caused by noise and fluctuation, and it is computationally efficient as compared to several other methods. Such properties are particularly suited for remote sensing classification problems and explain their recent adoption. In this poster, the SVM classification of ALS LiDAR data is proposed. Firstly, connected component analysis is applied for clustering the point cloud. Secondly, the resulting clusters are incorporated in the SVM classifier. Radial basic function (RFB) kernel is used due to the few numbers of parameters (C and γ) that needs to be chosen, which decreases the computation time. In order to optimize the classification rates, the parameters selection is explored. It consists to find the parameters (C and γ) leading to the best overall accuracy using grid search and 5-fold cross-validation. The exploited LiDAR point cloud is provided by the German Society for Photogrammetry, Remote Sensing, and Geoinformation. The ALS data used is characterized by a low density (4-6 points/m²) and is covering an urban area located in residential parts of the city Vaihingen in southern Germany. The class ground and three other classes belonging to roof superstructures are considered, i.e., a total of 4 classes. The training and test sets are selected randomly several times. The obtained results demonstrated that a parameters selection can orient the selection in a restricted interval of (C and γ) that can be further explored but does not systematically lead to the optimal rates. The SVM classifier with hyper-parameters is compared with the most used classifiers in literature for LiDAR data, random forest, AdaBoost, and decision tree. The comparison showed the superiority of the SVM classifier using parameters selection for LiDAR data compared to other classifiers.

Keywords: classification, airborne LiDAR, parameters selection, support vector machine

Procedia PDF Downloads 141
2258 Signal Integrity Performance Analysis in Capacitive and Inductively Coupled Very Large Scale Integration Interconnect Models

Authors: Mudavath Raju, Bhaskar Gugulothu, B. Rajendra Naik

Abstract:

The rapid advances in Very Large Scale Integration (VLSI) technology has resulted in the reduction of minimum feature size to sub-quarter microns and switching time in tens of picoseconds or even less. As a result, the degradation of high-speed digital circuits due to signal integrity issues such as coupling effects, clock feedthrough, crosstalk noise and delay uncertainty noise. Crosstalk noise in VLSI interconnects is a major concern and reduction in VLSI interconnect has become more important for high-speed digital circuits. It is the most effectively considered in Deep Sub Micron (DSM) and Ultra Deep Sub Micron (UDSM) technology. Increasing spacing in-between aggressor and victim line is one of the technique to reduce the crosstalk. Guard trace or shield insertion in-between aggressor and victim is also one of the prominent options for the minimization of crosstalk. In this paper, far end crosstalk noise is estimated with mutual inductance and capacitance RLC interconnect model. Also investigated the extent of crosstalk in capacitive and inductively coupled interconnects to minimizes the same through shield insertion technique.

Keywords: VLSI, interconnects, signal integrity, crosstalk, shield insertion, guard trace, deep sub micron

Procedia PDF Downloads 177
2257 Analysis Model for the Relationship of Users, Products, and Stores on Online Marketplace Based on Distributed Representation

Authors: Ke He, Wumaier Parezhati, Haruka Yamashita

Abstract:

Recently, online marketplaces in the e-commerce industry, such as Rakuten and Alibaba, have become some of the most popular online marketplaces in Asia. In these shopping websites, consumers can select purchase products from a large number of stores. Additionally, consumers of the e-commerce site have to register their name, age, gender, and other information in advance, to access their registered account. Therefore, establishing a method for analyzing consumer preferences from both the store and the product side is required. This study uses the Doc2Vec method, which has been studied in the field of natural language processing. Doc2Vec has been used in many cases to analyze the extraction of semantic relationships between documents (represented as consumers) and words (represented as products) in the field of document classification. This concept is applicable to represent the relationship between users and items; however, the problem is that one more factor (i.e., shops) needs to be considered in Doc2Vec. More precisely, a method for analyzing the relationship between consumers, stores, and products is required. The purpose of our study is to combine the analysis of the Doc2vec model for users and shops, and for users and items in the same feature space. This method enables the calculation of similar shops and items for each user. In this study, we derive the real data analysis accumulated in the online marketplace and demonstrate the efficiency of the proposal.

Keywords: Doc2Vec, online marketplace, marketing, recommendation systems

Procedia PDF Downloads 104
2256 A Systematic Review on Lifelong Learning Programs for Community-Dwelling Older Adults

Authors: Xi Vivien Wu, Emily Neo Kim Ang, Yi Jung Tung, Wenru Wang

Abstract:

Background and Objective: The increase in life expectancy and emphasis on self-reliance for the older adults are global phenomena. As such, lifelong learning in the community is considered a viable means of promoting successful and active aging. This systematic review aims to examine various lifelong learning programs for community-dwelling older adults and to synthesize the contents and outcomes of these lifelong learning programs. Methods: A systematic search was conducted in July to December 2016. Two reviewers were engaged in the process to ensure creditability of the selection process. Narrative description and analysis were applied with the support of a tabulation of key data including study design, interventions, and outcomes. Results: Eleven articles, which consisted of five randomized controlled trials and six quasi-experimental studies, were included in this review. Interventions included e-health literacy programs with the aid of computers and the Internet (n=4), computer and Internet training (n=3), physical fitness programs (n=2), music program (n=1), and intergenerational program (n=1). All studies used objective measurement tools to evaluate the outcomes of the study. Conclusion: The systematic review indicated lifelong learning programs resulted in positive outcomes in terms of physical health, mental health, social behavior, social support, self-efficacy and confidence in computer usage, and increased e-health literacy efficacy. However, the lifelong learning programs face challenges such as funding shortages, program cuts, and increasing costs. A comprehensive lifelong learning program could be developed to enhance the well-being of the older adults at a more holistic level. Empirical research can be done to explore the effectiveness of this comprehensive lifelong learning program.

Keywords: community-dwelling older adults, e-health literacy program, lifelong learning program, the wellbeing of the older adults

Procedia PDF Downloads 154
2255 Predicting Stack Overflow Accepted Answers Using Features and Models with Varying Degrees of Complexity

Authors: Osayande Pascal Omondiagbe, Sherlock a Licorish

Abstract:

Stack Overflow is a popular community question and answer portal which is used by practitioners to solve technology-related challenges during software development. Previous studies have shown that this forum is becoming a substitute for official software programming languages documentation. While tools have looked to aid developers by presenting interfaces to explore Stack Overflow, developers often face challenges searching through many possible answers to their questions, and this extends the development time. To this end, researchers have provided ways of predicting acceptable Stack Overflow answers by using various modeling techniques. However, less interest is dedicated to examining the performance and quality of typically used modeling methods, and especially in relation to models’ and features’ complexity. Such insights could be of practical significance to the many practitioners that use Stack Overflow. This study examines the performance and quality of various modeling methods that are used for predicting acceptable answers on Stack Overflow, drawn from 2014, 2015 and 2016. Our findings reveal significant differences in models’ performance and quality given the type of features and complexity of models used. Researchers examining classifiers’ performance and quality and features’ complexity may leverage these findings in selecting suitable techniques when developing prediction models.

Keywords: feature selection, modeling and prediction, neural network, random forest, stack overflow

Procedia PDF Downloads 123
2254 Facial Emotion Recognition with Convolutional Neural Network Based Architecture

Authors: Koray U. Erbas

Abstract:

Neural networks are appealing for many applications since they are able to learn complex non-linear relationships between input and output data. As the number of neurons and layers in a neural network increase, it is possible to represent more complex relationships with automatically extracted features. Nowadays Deep Neural Networks (DNNs) are widely used in Computer Vision problems such as; classification, object detection, segmentation image editing etc. In this work, Facial Emotion Recognition task is performed by proposed Convolutional Neural Network (CNN)-based DNN architecture using FER2013 Dataset. Moreover, the effects of different hyperparameters (activation function, kernel size, initializer, batch size and network size) are investigated and ablation study results for Pooling Layer, Dropout and Batch Normalization are presented.

Keywords: convolutional neural network, deep learning, deep learning based FER, facial emotion recognition

Procedia PDF Downloads 249