Search results for: threshold graphs
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1064

Search results for: threshold graphs

494 Weighted-Distance Sliding Windows and Cooccurrence Graphs for Supporting Entity-Relationship Discovery in Unstructured Text

Authors: Paolo Fantozzi, Luigi Laura, Umberto Nanni

Abstract:

The problem of Entity relation discovery in structured data, a well covered topic in literature, consists in searching within unstructured sources (typically, text) in order to find connections among entities. These can be a whole dictionary, or a specific collection of named items. In many cases machine learning and/or text mining techniques are used for this goal. These approaches might be unfeasible in computationally challenging problems, such as processing massive data streams. A faster approach consists in collecting the cooccurrences of any two words (entities) in order to create a graph of relations - a cooccurrence graph. Indeed each cooccurrence highlights some grade of semantic correlation between the words because it is more common to have related words close each other than having them in the opposite sides of the text. Some authors have used sliding windows for such problem: they count all the occurrences within a sliding windows running over the whole text. In this paper we generalise such technique, coming up to a Weighted-Distance Sliding Window, where each occurrence of two named items within the window is accounted with a weight depending on the distance between items: a closer distance implies a stronger evidence of a relationship. We develop an experiment in order to support this intuition, by applying this technique to a data set consisting in the text of the Bible, split into verses.

Keywords: cooccurrence graph, entity relation graph, unstructured text, weighted distance

Procedia PDF Downloads 143
493 Evaluating the Diagnostic Accuracy of the ctDNA Methylation for Liver Cancer

Authors: Maomao Cao

Abstract:

Objective: To test the performance of ctDNA methylation for the detection of liver cancer. Methods: A total of 1233 individuals have been recruited in 2017. 15 male and 15 female samples (including 10 cases of liver cancer) were randomly selected in the present study. CfDNA was extracted by MagPure Circulating DNA Maxi Kit. The concentration of cfDNA was obtained by Qubit™ dsDNA HS Assay Kit. A pre-constructed predictive model was used to analyze methylation data and to give a predictive score for each cfDNA sample. Individuals with a predictive score greater than or equal to 80 were classified as having liver cancer. CT tests were considered the gold standard. Sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for the diagnosis of liver cancer were calculated. Results: 9 patients were diagnosed with liver cancer according to the prediction model (with high sensitivity and threshold of 80 points), with scores of 99.2, 91.9, 96.6, 92.4, 91.3, 92.5, 96.8, 91.1, and 92.2, respectively. The sensitivity, specificity, positive predictive value, and negative predictive value of ctDNA methylation for the diagnosis of liver cancer were 0.70, 0.90, 0.78, and 0.86, respectively. Conclusions: ctDNA methylation could be an acceptable diagnostic modality for the detection of liver cancer.

Keywords: liver cancer, ctDNA methylation, detection, diagnostic performance

Procedia PDF Downloads 146
492 Influence of Measurement System on Negative Bias Temperature Instability Characterization: Fast BTI vs Conventional BTI vs Fast Wafer Level Reliability

Authors: Vincent King Soon Wong, Hong Seng Ng, Florinna Sim

Abstract:

Negative Bias Temperature Instability (NBTI) is one of the critical degradation mechanisms in semiconductor device reliability that causes shift in the threshold voltage (Vth). However, thorough understanding of this reliability failure mechanism is still unachievable due to a recovery characteristic known as NBTI recovery. This paper will demonstrate the severity of NBTI recovery as well as one of the effective methods used to mitigate, which is the minimization of measurement system delays. Comparison was done in between two measurement systems that have significant differences in measurement delays to show how NBTI recovery causes result deviations and how fast measurement systems can mitigate NBTI recovery. Another method to minimize NBTI recovery without the influence of measurement system known as Fast Wafer Level Reliability (FWLR) NBTI was also done to be used as reference.

Keywords: fast vs slow BTI, fast wafer level reliability (FWLR), negative bias temperature instability (NBTI), NBTI measurement system, metal-oxide-semiconductor field-effect transistor (MOSFET), NBTI recovery, reliability

Procedia PDF Downloads 417
491 Cloud Data Security Using Map/Reduce Implementation of Secret Sharing Schemes

Authors: Sara Ibn El Ahrache, Tajje-eddine Rachidi, Hassan Badir, Abderrahmane Sbihi

Abstract:

Recently, there has been increasing confidence for a favorable usage of big data drawn out from the huge amount of information deposited in a cloud computing system. Data kept on such systems can be retrieved through the network at the user’s convenience. However, the data that users send include private information, and therefore, information leakage from these data is now a major social problem. The usage of secret sharing schemes for cloud computing have lately been approved to be relevant in which users deal out their data to several servers. Notably, in a (k,n) threshold scheme, data security is assured if and only if all through the whole life of the secret the opponent cannot compromise more than k of the n servers. In fact, a number of secret sharing algorithms have been suggested to deal with these security issues. In this paper, we present a Mapreduce implementation of Shamir’s secret sharing scheme to increase its performance and to achieve optimal security for cloud data. Different tests were run and through it has been demonstrated the contributions of the proposed approach. These contributions are quite considerable in terms of both security and performance.

Keywords: cloud computing, data security, Mapreduce, Shamir's secret sharing

Procedia PDF Downloads 299
490 A Critical Appraisal of CO₂ Entrance Pressure with Heat

Authors: Abrar Al-Mutairi, Talal Al-Bazali

Abstract:

In this study, changes in capillary entry pressure of shale, as it interacts with CO₂, under different temperatures (25 °C to 250 °C) have been investigated. The combined impact of temperature and petrophysical properties (water content, water activity, permeability and porosity) of shale was also addressed. Results showed that the capillary entry pressure of shale when it interacted with CO₂ was highly affected by temperature. In general, increasing the temperature decreased capillary entry pressure of shale. We believe that pore dilation, where pore throat size expands due to the application of heat, may have caused this decrease in capillary entry pressure of shale. However, in some cases we found that at higher temperature some shale samples showed that the temperature activated clay swelling may have caused an apparent decrease in pore throat radii of shale which translates into higher capillary entry pressure of shale. Also, our results showed that there is no distinct relationship between shale’s water content, water activity, permeability, and porosity on the capillary entry pressure of shale samples as it interacted with CO₂ at different temperatures.

Keywords: heat, threshold pressure, CO₂ sequestration, shale

Procedia PDF Downloads 110
489 Performance of Slot-Entry Hybrid Worn Journal Bearing under Turbulent Lubrication

Authors: Nathi Ram, Saurabh K. Yadav

Abstract:

In turbomachinery, the turbulent flow occurs due to the use of high velocity of low kinematic viscosity lubricants and used in many industrial applications. In the present work, the performance of symmetric slot-entry hybrid worn journal bearing under laminar and turbulent lubrication has been investigated. For turbulent lubrication, the Reynolds equation has been modified using Constantinescu turbulent model. This modified equation has been solved using the finite element method. The effect of turbulent lubrication on bearing’s performance has been presented for symmetric hybrid journal bearing. The slot-entry hybrid worn journal bearing under turbulent/laminar regimes have been investigated. It has been observed that the stiffness and damping coefficients are more for the bearing having slot width ratio (SWR) of 0.25 than the bearing with SWR of 0.5 and 0.75 under the turbulent regime. Further, it is also observed that for constant wear depth parameter, stability threshold speed gets increased for bearing operates at slot width ratio 0.25 under turbulent lubrication.

Keywords: hydrostatic bearings, journal bearings, restrictors, turbulent flow models, finite element technique

Procedia PDF Downloads 158
488 Facility Anomaly Detection with Gaussian Mixture Model

Authors: Sunghoon Park, Hank Kim, Jinwon An, Sungzoon Cho

Abstract:

Internet of Things allows one to collect data from facilities which are then used to monitor them and even predict malfunctions in advance. Conventional quality control methods focus on setting a normal range on a sensor value defined between a lower control limit and an upper control limit, and declaring as an anomaly anything falling outside it. However, interactions among sensor values are ignored, thus leading to suboptimal performance. We propose a multivariate approach which takes into account many sensor values at the same time. In particular Gaussian Mixture Model is used which is trained to maximize likelihood value using Expectation-Maximization algorithm. The number of Gaussian component distributions is determined by Bayesian Information Criterion. The negative Log likelihood value is used as an anomaly score. The actual usage scenario goes like a following. For each instance of sensor values from a facility, an anomaly score is computed. If it is larger than a threshold, an alarm will go off and a human expert intervenes and checks the system. A real world data from Building energy system was used to test the model.

Keywords: facility anomaly detection, gaussian mixture model, anomaly score, expectation maximization algorithm

Procedia PDF Downloads 267
487 Detection of Patient Roll-Over Using High-Sensitivity Pressure Sensors

Authors: Keita Nishio, Takashi Kaburagi, Yosuke Kurihara

Abstract:

Recent advances in medical technology have served to enhance average life expectancy. However, the total time for which the patients are prescribed complete bedrest has also increased. With patients being required to maintain a constant lying posture- also called bedsore- development of a system to detect patient roll-over becomes imperative. For this purpose, extant studies have proposed the use of cameras, and favorable results have been reported. Continuous on-camera monitoring, however, tends to violate patient privacy. We have proposed unconstrained bio-signal measurement system that could detect body-motion during sleep and does not violate patient’s privacy. Therefore, in this study, we propose a roll-over detection method by the date obtained from the bi-signal measurement system. Signals recorded by the sensor were assumed to comprise respiration, pulse, body motion, and noise components. Compared the body-motion and respiration, pulse component, the body-motion, during roll-over, generate large vibration. Thus, analysis of the body-motion component facilitates detection of the roll-over tendency. The large vibration associated with the roll-over motion has a great effect on the Root Mean Square (RMS) value of time series of the body motion component calculated during short 10 s segments. After calculation, the RMS value during each segment was compared to a threshold value set in advance. If RMS value in any segment exceeded the threshold, corresponding data were considered to indicate occurrence of a roll-over. In order to validate the proposed method, we conducted experiment. A bi-directional microphone was adopted as a high-sensitivity pressure sensor and was placed between the mattress and bedframe. Recorded signals passed through an analog Band-pass Filter (BPF) operating over the 0.16-16 Hz bandwidth. BPF allowed the respiration, pulse, and body-motion to pass whilst removing the noise component. Output from BPF was A/D converted with the sampling frequency 100Hz, and the measurement time was 480 seconds. The number of subjects and data corresponded to 5 and 10, respectively. Subjects laid on a mattress in the supine position. During data measurement, subjects—upon the investigator's instruction—were asked to roll over into four different positions—supine to left lateral, left lateral to prone, prone to right lateral, and right lateral to supine. Recorded data was divided into 48 segments with 10 s intervals, and the corresponding RMS value for each segment was calculated. The system was evaluated by the accuracy between the investigator’s instruction and the detected segment. As the result, an accuracy of 100% was achieved. While reviewing the time series of recorded data, segments indicating roll-over tendencies were observed to demonstrate a large amplitude. However, clear differences between decubitus and the roll-over motion could not be confirmed. Extant researches possessed a disadvantage in terms of patient privacy. The proposed study, however, demonstrates more precise detection of patient roll-over tendencies without violating their privacy. As a future prospect, decubitus estimation before and after roll-over could be attempted. Since in this paper, we could not confirm the clear differences between decubitus and the roll-over motion, future studies could be based on utilization of the respiration and pulse components.

Keywords: bedsore, high-sensitivity pressure sensor, roll-over, unconstrained bio-signal measurement

Procedia PDF Downloads 115
486 Numerical Solution of Space Fractional Order Linear/Nonlinear Reaction-Advection Diffusion Equation Using Jacobi Polynomial

Authors: Shubham Jaiswal

Abstract:

During modelling of many physical problems and engineering processes, fractional calculus plays an important role. Those are greatly described by fractional differential equations (FDEs). So a reliable and efficient technique to solve such types of FDEs is needed. In this article, a numerical solution of a class of fractional differential equations namely space fractional order reaction-advection dispersion equations subject to initial and boundary conditions is derived. In the proposed approach shifted Jacobi polynomials are used to approximate the solutions together with shifted Jacobi operational matrix of fractional order and spectral collocation method. The main advantage of this approach is that it converts such problems in the systems of algebraic equations which are easier to be solved. The proposed approach is effective to solve the linear as well as non-linear FDEs. To show the reliability, validity and high accuracy of proposed approach, the numerical results of some illustrative examples are reported, which are compared with the existing analytical results already reported in the literature. The error analysis for each case exhibited through graphs and tables confirms the exponential convergence rate of the proposed method.

Keywords: space fractional order linear/nonlinear reaction-advection diffusion equation, shifted Jacobi polynomials, operational matrix, collocation method, Caputo derivative

Procedia PDF Downloads 440
485 A Near-Optimal Domain Independent Approach for Detecting Approximate Duplicates

Authors: Abdelaziz Fellah, Allaoua Maamir

Abstract:

We propose a domain-independent merging-cluster filter approach complemented with a set of algorithms for identifying approximate duplicate entities efficiently and accurately within a single and across multiple data sources. The near-optimal merging-cluster filter (MCF) approach is based on the Monge-Elkan well-tuned algorithm and extended with an affine variant of the Smith-Waterman similarity measure. Then we present constant, variable, and function threshold algorithms that work conceptually in a divide-merge filtering fashion for detecting near duplicates as hierarchical clusters along with their corresponding representatives. The algorithms take recursive refinement approaches in the spirit of filtering, merging, and updating, cluster representatives to detect approximate duplicates at each level of the cluster tree. Experiments show a high effectiveness and accuracy of the MCF approach in detecting approximate duplicates by outperforming the seminal Monge-Elkan’s algorithm on several real-world benchmarks and generated datasets.

Keywords: data mining, data cleaning, approximate duplicates, near-duplicates detection, data mining applications and discovery

Procedia PDF Downloads 381
484 Articulating Competencies Confidently: Employability in the Curriculum

Authors: Chris Procter

Abstract:

There is a significant debate on the role of University education in developing or teaching employability skills. Should higher education attempt to do this? Is it the best place? Is it able to do so? Different views abound, but the question is wrongly posed – one of the reasons that previous employability initiatives foundered (e.g., in the UK). Our role is less to teach than to guide, less to develop and more to help articulate: “the mind is not a vessel to be filled, but a fire to be lit” (Plutarch). This paper then addresses how this can be achieved taking into account criticism of employability initiatives as well as relevant learning theory. It discusses the experience of a large module which involved students being assessed on all stages of application for a live job description together with reflection on their professional development. The assessment itself adopted a Patchwork Text approach as a vehicle for learning. Students were guided to evaluate their strengths and areas to be developed, articulate their competencies, and reflect upon their development, moving on to new Thresholds of Employability. The paper uses the student voices to express the progress they made. It concludes that employability can and should be an effective part of the higher education curriculum when designed to encourage students to confidently articulate their competencies and take charge of their own professional development.

Keywords: competencies, employability, patchwork assessment, threshold concepts

Procedia PDF Downloads 210
483 Minimizing the Impact of Covariate Detection Limit in Logistic Regression

Authors: Shahadut Hossain, Jacek Wesolowski, Zahirul Hoque

Abstract:

In many epidemiological and environmental studies covariate measurements are subject to the detection limit. In most applications, covariate measurements are usually truncated from below which is known as left-truncation. Because the measuring device, which we use to measure the covariate, fails to detect values falling below the certain threshold. In regression analyses, it causes inflated bias and inaccurate mean squared error (MSE) to the estimators. This paper suggests a response-based regression calibration method to correct the deleterious impact introduced by the covariate detection limit in the estimators of the parameters of simple logistic regression model. Compared to the maximum likelihood method, the proposed method is computationally simpler, and hence easier to implement. It is robust to the violation of distributional assumption about the covariate of interest. In producing correct inference, the performance of the proposed method compared to the other competing methods has been investigated through extensive simulations. A real-life application of the method is also shown using data from a population-based case-control study of non-Hodgkin lymphoma.

Keywords: environmental exposure, detection limit, left truncation, bias, ad-hoc substitution

Procedia PDF Downloads 229
482 Assessing the Bioactivity and Cell Viability of Apatite-Wollastonite Glass Ceramics Prepared via Spray Pyrolysis

Authors: Andualem Workie

Abstract:

In this study, we examined the sinterability and bioactivity of MgO-SiO₂-P₂O₅-CaO-CaF₂ glass compositions created through spray pyrolysis. We evaluated the bioactivity of the materials by immersing them for varying periods of time in simulated bodily fluid (SBF) and found that bioactivity was related to the sintering temperature and soaking time. The material's pH value during immersion in SBF was within the range of 7.4-8.2, which is below 8.5 and improves compatibility and reduces toxicity in biological applications. We used X-ray diffraction and scanning electron microscopy to determine the phase compositions and morphologies of the samples and found that the 1100°C sintered A-W GC sample exhibited the highest bioactivity after soaking in SBF. This sample was dominated by fluorapatite, wollastonite, and whitlockite crystals scattered throughout the glass matrix. The crystallinity (%) of the A-W GC increased as its bioactivity improved, making it more suitable for use in pharmaceutical applications. We also conducted a cytotoxicity test on A-W GC samples sintered at different temperatures and found that the glass-ceramics were non-toxic to MC3T3-E1 cells at all extraction concentrations, except for those sintered at 700°C at concentrations of 250, 200, and 150 mg/ml where cell viability (%) was below the threshold of 70%.

Keywords: apatite wollastonite glass ceramics, bioactivity, calcination, cell viability

Procedia PDF Downloads 98
481 Simulation of Binary Nitride Inclusions Effect on Tensile Properties of Steel

Authors: Ali Dalirbod, Peyman Ahmadian

Abstract:

Inclusions are unavoidable part of all steels. Non-metallic inclusions have significant effects on mechanical properties of steel. The effects of inclusion on stress concentration around the matrix/inclusion have been extensively studied. The results relating to single inclusion behavior, describe properly the behavior of stress but not the elongation drop. The raised stress in inclusion/matrix results in crack initiation. The influence of binary inclusions on stress concentration around matrix is a major aim of this work which is representative of the simple pattern distribution of non-metallic inclusions. Stress concentration around inclusions in this case depends on parameters like distance between two inclusions (d), angle between centrally linking line of two inclusions, load axis (φ), and rotational angle of inclusion (θ). FEM analysis was applied to investigate the highest and lowest ductility versus varying parameters above. The simulation results show that there is a critical distance between two cubic inclusions in which bigger than the threshold, the stress, and strain field in matrix/inclusions interface converts into individual fields around each inclusion.

Keywords: nitride inclusion, simulation, tensile properties, inclusion-matrix interface

Procedia PDF Downloads 313
480 An Investigation on Orthopedic Rehabilitation by Avoiding Thermal Necrosis

Authors: R. V. Dahibhate, A. B. Deoghare, P. M. Padole

Abstract:

Maintaining natural integrity of biosystem is paramount significant for orthopedic surgeon while performing surgery. Restoration is challenging task to rehabilitate trauma patient. Drilling is an inevitable procedure to fix implants. The task leads to rise in temperature at the contact site which intends to thermal necrosis. A precise monitoring can avoid thermal necrosis. To accomplish it, data acquiring instrument is integrated with the drill bit. To contemplate it, electronic feedback system is developed. It not only measures temperature without any physical contact in between measuring device and target but also visualizes the site and monitors correct movement of tool path. In the current research work an infrared thermometer data acquisition system is used which monitors variation in temperature at the drilling site and a camera captured movement of drill bit advancement. The result is presented in graphical form which represents variations in temperature, drill rotation and time. A feedback system helps in keeping drill speed in threshold limit.

Keywords: thermal necrosis, infrared thermometer, drilling tool, feedback system

Procedia PDF Downloads 227
479 Determinants of Extra Charges for Container Shipments: A Case Study of Nexus Zone Logistics

Authors: Zety Shakila Binti Mohd Yusof, Muhammad Adib Bin Ishak, Hajah Fatimah Binti Hussein

Abstract:

The international shipping business is related to numerous controls or regulations of export and import shipments. It is costly and time consuming, and when something goes wrong or when the buyer or seller fails to comply with the regulations, it can result in penalties, delays, and unexpected costs etc. For the focus of this study, the researchers have selected a local forwarder that provides forwarding and clearance services, Nexus Zone Logistics. It was identified that this company currently has many extra costs to be paid including local and detention charges, which negatively impacts the flow of income and reduces overall stability. Two variables have been identified as factors of extra charges; loaded containers entering the port by exceeded closing time and late delivery of empty containers to the container yard. This study is a qualitative in nature and the secondary data collected was analyzed using self-administered observation. The findings of this study were covered by one selected case for each export and import shipment between July and December 2014. The data were analyzed using frequency analysis based on tables and graphs. The researcher recommends Nexus Zone Logistics impose a 1% deposit payment per container for each shipment (export and import) to its customers.

Keywords: international shipping, export and import, detention charges, container shipment

Procedia PDF Downloads 377
478 Comparative Study of Non-Identical Firearms with Priority to Repair Subject to Inspection

Authors: A. S. Grewal, R. S. Sangwan, Dharambir, Vikas Dhanda

Abstract:

The purpose of this paper is to develop and analyze two reliability models for a system of non-identical firearms – one is standard firearm (called as original unit) and the other is a country-made firearm (called as duplicate /substandard unit). There is a single server who comes immediately to do inspection and repair whenever needed. On the failure of standard firearm, the server inspects the operative country-made firearm to see whether the unit is capable of performing the desired function well or not. If country-made firearm is not capable to do so, the operation of the system is stopped and server starts repair of the standard firearms immediately. However, no inspection is done at the failure of the country-made firearm as the country-made firearm alone is capable of performing the given task well. In model I, priority to repair the standard firearm is given in case system fails completely and country-made firearm is already under repair, whereas in model II there is no such priority. The failure and repair times of each unit are assumed to be independent and uncorrelated random variables. The distributions of failure time of the units are taken as negative exponential while that of repair and inspection times are general. By using semi-Markov process and regenerative point technique some econo-reliability measures are obtained. Graphs are plotted to compare the MTSF (mean time to system failure), availability and profit of the models for a particular case.

Keywords: non-identical firearms, inspection, priority to repair, semi-Markov process, regenerative point

Procedia PDF Downloads 420
477 A Novel Software Model for Enhancement of System Performance and Security through an Optimal Placement of PMU and FACTS

Authors: R. Kiran, B. R. Lakshmikantha, R. V. Parimala

Abstract:

Secure operation of power systems requires monitoring of the system operating conditions. Phasor measurement units (PMU) are the device, which uses synchronized signals from the GPS satellites, and provide the phasors information of voltage and currents at a given substation. The optimal locations for the PMUs must be determined, in order to avoid redundant use of PMUs. The objective of this paper is to make system observable by using minimum number of PMUs & the implementation of stability software at 22OkV grid for on-line estimation of the power system transfer capability based on voltage and thermal limitations and for security monitoring. This software utilizes State Estimator (SE) and synchrophasor PMU data sets for determining the power system operational margin under normal and contingency conditions. This software improves security of transmission system by continuously monitoring operational margin expressed in MW or in bus voltage angles, and alarms the operator if the margin violates a pre-defined threshold.

Keywords: state estimator (SE), flexible ac transmission systems (FACTS), optimal location, phasor measurement units (PMU)

Procedia PDF Downloads 405
476 Multi-Wavelength Q-Switched Erbium-Doped Fiber Laser with Photonic Crystal Fiber and Multi-Walled Carbon Nanotubes

Authors: Zian Cheak Tiu, Harith Ahmad, Sulaiman Wadi Harun

Abstract:

A simple multi-wavelength passively Q-switched Erbium-doped fiber laser (EDFL) is demonstrated using low cost multi-walled carbon nanotubes (MWCNTs) based saturable absorber (SA), which is prepared using polyvinyl alcohol (PVA) as a host polymer. The multi-wavelength operation is achieved based on nonlinear polarization rotation (NPR) effect by incorporating 50 m long photonic crystal fiber (PCF) in the ring cavity. The EDFL produces a stable multi-wavelength comb spectrum for more than 14 lines with a fixed spacing of 0.48 nm. The laser also demonstrates a stable pulse train with the repetition rate increases from 14.9 kHz to 25.4 kHz as the pump power increases from the threshold power of 69.0 mW to the maximum pump power of 133.8 mW. The minimum pulse width of 4.4 µs was obtained at the maximum pump power of 133.8 mW while the highest energy of 0.74 nJ was obtained at pump power of 69.0 mW.

Keywords: multi-wavelength Q-switched, multi-walled carbon nanotube, photonic crystal fiber

Procedia PDF Downloads 530
475 Reliability Factors Based Fuzzy Logic Scheme for Spectrum Sensing

Authors: Tallataf Rasheed, Adnan Rashdi, Ahmad Naeem Akhtar

Abstract:

The accurate spectrum sensing is a fundamental requirement of dynamic spectrum access for deployment of Cognitive Radio Network (CRN). To acheive this requirement a Reliability factors based Fuzzy Logic (RFL) Scheme for Spectrum Sensing has been proposed in this paper. Cognitive Radio User (CRU) predicts the presence or absence of Primary User (PU) using energy detector and calculates the Reliability factors which are SNR of sensing node, threshold of energy detector and decision difference of each node with other nodes in a cooperative spectrum sensing environment. Then the decision of energy detector is combined with Reliability factors of sensing node using Fuzzy Logic. These Reliability Factors used in RFL Scheme describes the reliability of decision made by a CRU to improve the local spectrum sensing. This Fuzzy combining scheme provides the accuracy of decision made by sensornode. The simulation results have shown that the proposed technique provide better PU detection probability than existing Spectrum Sensing Techniques.

Keywords: cognitive radio, spectrum sensing, energy detector, reliability factors, fuzzy logic

Procedia PDF Downloads 479
474 Decision Support: How Explainable A.I. Can Improve Transparency and Trust with Human Users

Authors: Devon Brown, Liu Chunmei

Abstract:

This paper will present an analysis as part of the researchers dissertation topic focusing on the intersection of affective and analytical directed acyclic graphs (DAGs) in the context of Decision Support Systems (DSS). The researcher’s work involves analyzing decision theory models like Affective and Bayesian Decision theory models and how they could be implemented under an Affective Computing Framework using Information Fusion and Human-Centered Design. Additionally, the researcher is beginning research on an Affective-Analytic Decision Framework (AADF) model for their dissertation research and are looking to merge logic and analytic models with empathetic insights into affective DAGs. Data-collection efforts begin Fall 2024 and in preparation for the efforts this paper looks to analyze previous research in this area and introduce the AADF framework and propose conceptual models for consideration. For this paper, the research emphasis is placed on analyzing Bayesian networks and Markov models which offer probabilistic techniques during uncertainty in decision-making. Ideally, including affect into analytic models will ensure algorithms can increase user trust with algorithms by including emotional states and the user’s experience with the goal of developing emotionally intelligent A.I. systems that can start to navigate the complex fabric of human emotion during decision-making.

Keywords: decision support systems, explainable AI, HCAI techniques, affective-analytical decision framework

Procedia PDF Downloads 8
473 A Framework for Chinese Domain-Specific Distant Supervised Named Entity Recognition

Authors: Qin Long, Li Xiaoge

Abstract:

The Knowledge Graphs have now become a new form of knowledge representation. However, there is no consensus in regard to a plausible and definition of entities and relationships in the domain-specific knowledge graph. Further, in conjunction with several limitations and deficiencies, various domain-specific entities and relationships recognition approaches are far from perfect. Specifically, named entity recognition in Chinese domain is a critical task for the natural language process applications. However, a bottleneck problem with Chinese named entity recognition in new domains is the lack of annotated data. To address this challenge, a domain distant supervised named entity recognition framework is proposed. The framework is divided into two stages: first, the distant supervised corpus is generated based on the entity linking model of graph attention neural network; secondly, the generated corpus is trained as the input of the distant supervised named entity recognition model to train to obtain named entities. The link model is verified in the ccks2019 entity link corpus, and the F1 value is 2% higher than that of the benchmark method. The re-pre-trained BERT language model is added to the benchmark method, and the results show that it is more suitable for distant supervised named entity recognition tasks. Finally, it is applied in the computer field, and the results show that this framework can obtain domain named entities.

Keywords: distant named entity recognition, entity linking, knowledge graph, graph attention neural network

Procedia PDF Downloads 87
472 Convex Restrictions for Outage Constrained MU-MISO Downlink under Imperfect Channel State Information

Authors: A. Preetha Priyadharshini, S. B. M. Priya

Abstract:

In this paper, we consider the MU-MISO downlink scenario, under imperfect channel state information (CSI). The main issue in imperfect CSI is to keep the probability of each user achievable outage rate below the given threshold level. Such a rate outage constraints present significant and analytical challenges. There are many probabilistic methods are used to minimize the transmit optimization problem under imperfect CSI. Here, decomposition based large deviation inequality and Bernstein type inequality convex restriction methods are used to perform the optimization problem under imperfect CSI. These methods are used for achieving improved output quality and lower complexity. They provide a safe tractable approximation of the original rate outage constraints. Based on these method implementations, performance has been evaluated in the terms of feasible rate and average transmission power. The simulation results are shown that all the two methods offer significantly improved outage quality and lower computational complexity.

Keywords: imperfect channel state information, outage probability, multiuser- multi input single output, channel state information

Procedia PDF Downloads 810
471 The Use of Culture as a Campaign Method in Indonesian Parliamentary Election

Authors: Azza Habibullah

Abstract:

The principal objective of this paper is to show the use of participatory culture in the parliamentarian campaign. The use of this method has always been non-popular amongst the parliamentarian candidates due to the amount of times and energy that they need to spent with the constituents. However, due to many parliamentarian corruption cases in the last five years period, some political party have been losing peoples trust. That political party trust lost had also affecting the parliamentarian candidates electability, so they invent some creative campaign method that involving their constituent with more intimates and friendly environment. In this paper, an observation is done to a parliamentarian candidate from Partai Keadilan Sejahtera (Prosperous Justice Party) in Bandung and Cimahi City area, West Java. This Parliamentraian candidate is known for her personal-approach campaign method such as a puppet show, hanging out with group of ex-bike gang leaders, and going fishing with the constituent. This paper will compare her method with other parliamentarian candidates from the same party as her that mostly use mainstream campaign method such as open speech, print media, an other one way campaign method. While the other parliamentarian candidates failed to reach the parliamentarian threshold, the participatory method had proven as an effective method.

Keywords: participatory culture, Indonesian parliamentary election, Prosperous Justice Party, electability

Procedia PDF Downloads 259
470 Catalytic Thermodynamics of Nanocluster Adsorbates from Informational Statistical Mechanics

Authors: Forrest Kaatz, Adhemar Bultheel

Abstract:

We use an informational statistical mechanics approach to study the catalytic thermodynamics of platinum and palladium cuboctahedral nanoclusters. Nanoclusters and their adatoms are viewed as chemical graphs with a nearest neighbor adjacency matrix. We use the Morse potential to determine bond energies between cluster atoms in a coordination type calculation. We use adsorbate energies calculated from density functional theory (DFT) to study the adatom effects on the thermodynamic quantities, which are derived from a Hamiltonian. Oxygen radical and molecular adsorbates are studied on platinum clusters and hydrogen on palladium clusters. We calculate the entropy, free energy, and total energy as the coverage of adsorbates increases from bridge and hollow sites on the surface. Thermodynamic behavior versus adatom coverage is related to the structural distribution of adatoms on the nanocluster surfaces. The thermodynamic functions are characterized using a simple adsorption model, with linear trends as the coverage of adatoms increases. The data exhibits size effects for the measured thermodynamic properties with cluster diameters between 2 and 5 nm. Entropy and enthalpy calculations of Pt-O2 compare well with previous theoretical data for Pt(111)-O2, and our Pd-H results show similar trends as experimental measurements for Pd-H2 nanoclusters. Our methods are general and may be applied to wide variety of nanocluster adsorbate systems.

Keywords: catalytic thermodynamics, palladium nanocluster absorbates, platinum nanocluster absorbates, statistical mechanics

Procedia PDF Downloads 158
469 Developing Cucurbitacin a Minimum Inhibition Concentration of Meloidogyne Incognita Using a Computer-Based Model

Authors: Zakheleni P. Dube, Phatu W. Mashela

Abstract:

Minimum inhibition concentration (MIC) is the lowest concentration of a chemical that brings about significant inhibition of target organism. The conventional method for establishing the MIC for phytonematicides is tedious. The objective of this study was to use the Curve-fitting Allelochemical Response Data (CARD) to determine the MIC for pure cucurbitacin A on Meloidogyne incognita second-stage juveniles (J2) hatch, immobility and mortality. Meloidogyne incognita eggs and freshly hatched J2 were separately exposed to a series of pure cucurbitacin A concentrations of 0.00, 0.25, 0.50, 0.75, 1.00, 1.25, 1.50, 1.75, 2.00, 2.25 and 2.50 μg.mL⁻¹for 12, 24, 48 and 72 h in an incubator set at 25 ± 2°C. Meloidogyne incognita J2 hatch, immobility and mortality counts were determined using a stereomicroscope and the significant means were subjected to the CARD model. The model exhibited density-dependent growth (DDG) patterns of J2 hatch, immobility and mortality to increasing concentrations of cucurbitacin A. The average MIC for cucurbitacin A on M. incognita J2 hatch, immobility and mortality were 2.2, 0.58 and 0.63 µg.mL⁻¹, respectively. Meloidogyne incognita J2 hatch had the highest average MIC value followed by mortality and immobility had the least. In conclusion, the CARD model was able to generate MIC for cucurbitacin A, hence it could serve as a valuable tool in the chemical-nematode bioassay studies.

Keywords: inhibition concentration, phytonematicide, sensitivity index, threshold stimulation, triterpenoids.

Procedia PDF Downloads 185
468 Development of Automatic Laser Scanning Measurement Instrument

Authors: Chien-Hung Liu, Yu-Fen Chen

Abstract:

This study used triangular laser probe and three-axial direction mobile platform for surface measurement, programmed it and applied it to real-time analytic statistics of different measured data. This structure was used to design a system integration program: using triangular laser probe for scattering or reflection non-contact measurement, transferring the captured signals to the computer through RS-232, and using RS-485 to control the three-axis platform for a wide range of measurement. The data captured by the laser probe are formed into a 3D surface. This study constructed an optical measurement application program in the concept of visual programming language. First, the signals are transmitted to the computer through RS-232/RS-485, and then the signals are stored and recorded in graphic interface timely. This programming concept analyzes various messages, and makes proper presentation graphs and data processing to provide the users with friendly graphic interfaces and data processing state monitoring, and identifies whether the present data are normal in graphic concept. The major functions of the measurement system developed by this study are thickness measurement, SPC, surface smoothness analysis, and analytical calculation of trend line. A result report can be made and printed promptly. This study measured different heights and surfaces successfully, performed on-line data analysis and processing effectively, and developed a man-machine interface for users to operate.

Keywords: laser probe, non-contact measurement, triangulation measurement principle, statistical process control, labVIEW

Procedia PDF Downloads 359
467 Design and Analysis of Metamaterial Based Vertical Cavity Surface Emitting Laser

Authors: Ishraq M. Anjum

Abstract:

Distributed Bragg reflectors are used in vertical-cavity surface-emitting lasers (VCSELs) in order to achieve very high reflectivity. Use of metamaterial in place of distributed Bragg reflector can reduce the device size significantly. A silicon-based metamaterial near perfect reflector is designed to be used in place of distributed Bragg reflectors in VCSELs. Mie resonance in dielectric microparticles is exploited in order to design the metamaterial. A reflectivity of 98.31% is achieved using finite-difference time-domain method. An 808nm double intra-cavity contacted VCSEL structure with 1.5 λ cavity is proposed using this metamaterial near perfect reflector. The active region is designed to be composed of seven GaAs/AlGaAs quantum wells. Upon numerical investigation of the designed VCSEL structure, the threshold current is found to be 2.96 mA at an aperture of 40 square micrometers and the maximum output power is found to be 71 mW at a current of 141 mA. Miniaturization of conventional VCSELs is possible using this design.

Keywords: GaAs, LASER, metamaterial, VCSEL, vertical cavity surface emitting laser

Procedia PDF Downloads 177
466 Modelling of Creep in a Thick-Walled Cylindrical Vessel Subjected to Internal Pressure

Authors: Tejeet Singh, Ishvneet Singh, Vinay Gupta

Abstract:

The present study focussed on carrying out the creep analysis in an isotropic thick-walled composite cylindrical pressure vessel composed of aluminium matrix reinforced with silicon-carbide in particulate form. The creep behaviour of the composite material has been described by the threshold stress based creep law. The value of stress exponent appearing in the creep law was selected as 3, 5 and 8. The constitutive equations were developed using well known von-Mises yield criteria. Models were developed to find out the distributions of creep stresses and strain rate in thick-walled composite cylindrical pressure vessels under internal pressure. In order to obtain the stress distributions in the cylinder, the equilibrium equation of the continuum mechanics and the constitutive equations are solved together. It was observed that the radial stress, tangential stress and axial stress increases along with the radial distance. The cross-over was also obtained almost at the middle region of cylindrical vessel for tangential and axial stress for different values of stress exponent. The strain rates were also decreasing in nature along the entire radius.

Keywords: creep, composite, cylindrical vessel, internal pressure

Procedia PDF Downloads 571
465 Easymodel: Web-based Bioinformatics Software for Protein Modeling Based on Modeller

Authors: Alireza Dantism

Abstract:

Presently, describing the function of a protein sequence is one of the most common problems in biology. Usually, this problem can be facilitated by studying the three-dimensional structure of proteins. In the absence of a protein structure, comparative modeling often provides a useful three-dimensional model of the protein that is dependent on at least one known protein structure. Comparative modeling predicts the three-dimensional structure of a given protein sequence (target) mainly based on its alignment with one or more proteins of known structure (templates). Comparative modeling consists of four main steps 1. Similarity between the target sequence and at least one known template structure 2. Alignment of target sequence and template(s) 3. Build a model based on alignment with the selected template(s). 4. Prediction of model errors 5. Optimization of the built model There are many computer programs and web servers that automate the comparative modeling process. One of the most important advantages of these servers is that it makes comparative modeling available to both experts and non-experts, and they can easily do their own modeling without the need for programming knowledge, but some other experts prefer using programming knowledge and do their modeling manually because by doing this they can maximize the accuracy of their modeling. In this study, a web-based tool has been designed to predict the tertiary structure of proteins using PHP and Python programming languages. This tool is called EasyModel. EasyModel can receive, according to the user's inputs, the desired unknown sequence (which we know as the target) in this study, the protein sequence file (template), etc., which also has a percentage of similarity with the primary sequence, and its third structure Predict the unknown sequence and present the results in the form of graphs and constructed protein files.

Keywords: structural bioinformatics, protein tertiary structure prediction, modeling, comparative modeling, modeller

Procedia PDF Downloads 90