Search results for: uniform error
956 Analytical Modelling of Surface Roughness during Compacted Graphite Iron Milling Using Ceramic Inserts
Authors: Ş. Karabulut, A. Güllü, A. Güldaş, R. Gürbüz
Abstract:
This study investigates the effects of the lead angle and chip thickness variation on surface roughness during the machining of compacted graphite iron using ceramic cutting tools under dry cutting conditions. Analytical models were developed for predicting the surface roughness values of the specimens after the face milling process. Experimental data was collected and imported to the artificial neural network model. A multilayer perceptron model was used with the back propagation algorithm employing the input parameters of lead angle, cutting speed and feed rate in connection with chip thickness. Furthermore, analysis of variance was employed to determine the effects of the cutting parameters on surface roughness. Artificial neural network and regression analysis were used to predict surface roughness. The values thus predicted were compared with the collected experimental data, and the corresponding percentage error was computed. Analysis results revealed that the lead angle is the dominant factor affecting surface roughness. Experimental results indicated an improvement in the surface roughness value with decreasing lead angle value from 88° to 45°.Keywords: CGI, milling, surface roughness, ANN, regression, modeling, analysis
Procedia PDF Downloads 448955 The Effect of Electromagnetic Stirring during Solidification of Nickel Based Alloys
Authors: Ricardo Paiva, Rui Soares, Felix Harnau, Bruno Fragoso
Abstract:
Nickel-based alloys are materials well suited for service in extreme environments subjected to pressure and heat. Some industrial applications for Nickel-based alloys are aerospace and jet engines, oil and gas extraction, pollution control and waste processing, automotive and marine industry. It is generally recognized that grain refinement is an effective methodology to improve the quality of casted parts. Conventional grain refinement techniques involve the addition of inoculation substances, the control of solidification conditions, or thermomechanical treatment with recrystallization. However, such methods often lead to non-uniform grain size distribution and the formation of hard phases, which are detrimental to both wear performance and biocompatibility. Stirring of the melt by electromagnetic fields has been widely used in continuous castings with success for grain refinement, solute redistribution, and surface quality improvement. Despite the advantages, much attention has not been paid yet to the use of this approach on functional castings such as investment casting. Furthermore, the effect of electromagnetic stirring (EMS) fields on Nickel-based alloys is not known. In line with the gaps/needs of the state-of-art, the present research work targets to promote new advances in controlling grain size and morphology of investment cast Nickel based alloys. For such a purpose, a set of experimental tests was conducted. A high-frequency induction furnace with vacuum and controlled atmosphere was used to cast the Inconel 718 alloy in ceramic shells. A coil surrounded the casting chamber in order to induce electromagnetic stirring during solidification. Aiming to assess the effect of the electromagnetic stirring on Ni alloys, the samples were subjected to microstructural analysis and mechanical tests. The results show that electromagnetic stirring can be an effective methodology to modify the grain size and mechanical properties of investment-cast parts.Keywords: investment casting, grain refinement, electromagnetic stirring, nickel alloys
Procedia PDF Downloads 132954 Capacity Oversizing for Infrastructure Sharing Synergies: A Game Theoretic Analysis
Authors: Robin Molinier
Abstract:
Industrial symbiosis (I.S) rely on two basic modes of cooperation between organizations that are infrastructure/service sharing and resource substitution (the use of waste materials, fatal energy and recirculated utilities for production). The former consists in the intensification of use of an asset and thus requires to compare the incremental investment cost to be incurred and the stand-alone cost faced by each potential participant to satisfy its own requirements. In order to investigate the way such a cooperation mode can be implemented we formulate a game theoretic model integrating the grassroot investment decision and the ex-post access pricing problem. In the first period two actors set cooperatively (resp. non-cooperatively) a level of common (resp. individual) infrastructure capacity oversizing to attract ex-post a potential entrant with a plug-and-play offer (available capacity, tariff). The entrant’s requirement is randomly distributed and known only after investments took place. Capacity cost exhibits sub-additive property so that there is room for profitable overcapacity setting in the first period under some conditions that we derive. The entrant willingness-to-pay for the access to the infrastructure is driven by both her standalone cost and the complement cost to be incurred in case she chooses to access an infrastructure whose the available capacity is lower than her requirement level. The expected complement cost function is thus derived, and we show that it is decreasing, convex and shaped by the entrant’s requirements distribution function. For both uniform and triangular distributions optimal capacity level is obtained in the cooperative setting and equilibrium levels are determined in the non-cooperative case. Regarding the latter, we show that competition is deterred by the first period investor with the highest requirement level. Using the non-cooperative game outcomes which gives lower bounds for the profit sharing problem in the cooperative one we solve the whole game and describe situations supporting sharing agreements.Keywords: capacity, cooperation, industrial symbiosis, pricing
Procedia PDF Downloads 440953 Performance Analysis of a Combined Ordered Successive and Interference Cancellation Using Zero-Forcing Detection over Rayleigh Fading Channels in Mimo Systems
Authors: Jamal R. Elbergali
Abstract:
Multiple Input Multiple Output (MIMO) systems are wireless systems with multiple antenna elements at both ends of the link. Wireless communication systems demand high data rate and spectral efficiency with increased reliability. MIMO systems have been popular techniques to achieve these goals because increased data rate is possible through spatial multiplexing scheme and diversity. Spatial Multiplexing (SM) is used to achieve higher possible throughput than diversity. In this paper, we propose a Zero-Forcing (ZF) detection using a combination of Ordered Successive Interference Cancellation (OSIC) and Zero Forcing using Interference Cancellation (ZF-IC). The proposed method used an OSIC based on Signal to Noise Ratio (SNR) ordering to get the estimation of last symbol (x ̃_(N_T )), then the estimated last symbol is considered to be an input to the ZF-IC. We analyze the Bit Error Rate (BER) performance of the proposed MIMO system over Rayleigh Fading Channel, using Binary Phase Shift Keying (BPSK) modulation scheme. The results show better performance than the previous methods.Keywords: SNR, BER, BPSK, MIMO, modulation, zero forcing (ZF), OSIC, ZF-IC, spatial multiplexing (SM)
Procedia PDF Downloads 423952 Spectrum Allocation Using Cognitive Radio in Wireless Mesh Networks
Authors: Ayoub Alsarhan, Ahmed Otoom, Yousef Kilani, Abdel-Rahman al-GHuwairi
Abstract:
Wireless mesh networks (WMNs) have emerged recently to improve internet access and other networking services. WMNs provide network access to the clients and other networking functions such as routing, and packet forwarding. Spectrum scarcity is the main challenge that limits the performance of WMNs. Cognitive radio is proposed to solve spectrum scarcity problem. In this paper, we consider a cognitive wireless mesh network where unlicensed users (secondary users, SUs) can access free spectrum that is allocated to spectrum owners (primary users, PUs). Although considerable research has been conducted on spectrum allocation, spectrum assignment is still considered an important challenging problem. This problem can be solved using cognitive radio technology that allows SUs to intelligently locate free bands and access them without interfering with PUs. Our scheme considers several heuristics for spectrum allocation. These heuristics include: channel error rate, PUs activities, channel capacity and channel switching time. Performance evaluation of the proposed scheme shows that the scheme is able to allocate the unused spectrum for SUs efficiently.Keywords: cognitive radio, dynamic spectrum access, spectrum management, spectrum sharing, wireless mesh networks
Procedia PDF Downloads 528951 Comparative Performance of Standing Whole Body Monitor and Shielded Chair Counter for In-vivo Measurements
Authors: M. Manohari, S. Priyadharshini, K. Bajeer Sulthan, R. Santhanam, S. Chandrasekaran, B. Venkatraman
Abstract:
In-vivo monitoring facility at Indira Gandhi Centre for Atomic Research (IGCAR), Kalpakkam, caters to the monitoring of internal exposure of occupational radiation workers from various radioactive facilities of IGCAR. Internal exposure measurement is done using Na(Tl) based Scintillation detectors. Two types of whole-body counters, namely Shielded Chair Counter (SC) and Standing Whole-Body Monitor (SWBM), are being used. The shielded Chair is based on a NaI detector of 20.3 cm diameter and 10.15 cm thick. The chair of the system is shielded using lead shots of 10 cm lead equivalent and the detector with 8 cm lead bricks. Counting geometry is sitting geometry. Calibration is done using 95 percentile BOMAB phantom. The minimum Detectable Activity (MDA) for 137Cs for the 60s is 1150 Bq. Standing Wholebody monitor (SWBM) has two NaI(Tl) detectors of size 10.16 x 10.16 x 40.64 cm3 positioned serially, one over the other. It has a shielding thickness of 5cm lead equivalent. Counting is done in standup geometry. Calibration is done with the help of Ortec Phantom, having a uniform distribution of mixed radionuclides for the thyroid, thorax and pelvis. The efficiency of SWBM is 2.4 to 3.5 times higher than that of the shielded chair in the energy range of 279 to 1332 keV. MDA of 250 Bq for 137Cs can be achieved with a counting time of 60s. MDA for 131I in the thyroid was estimated as 100 Bq from the MDA of whole-body for one-day post intake. Standing whole body monitor is better in terms of efficiency, MDA and ease of positioning. In case of emergency situations, the optimal MDAs for in-vivo monitoring service are 1000 Bq for 137Cs and 100 Bq for 131I. Hence, SWBM is more suitable for the rapid screening of workers as well as the public in the case of an emergency. While a person reports for counting, there is a potential for external contamination. In SWBM, there is a feasibility to discriminate them as the subject can be counted in anterior or posterior geometry which is not possible in SC.Keywords: minimum detectable activity, shielded chair, shielding thickness, standing whole body monitor
Procedia PDF Downloads 46950 Improve Closed Loop Performance and Control Signal Using Evolutionary Algorithms Based PID Controller
Authors: Mehdi Shahbazian, Alireza Aarabi, Mohsen Hadiyan
Abstract:
Proportional-Integral-Derivative (PID) controllers are the most widely used controllers in industry because of its simplicity and robustness. Different values of PID parameters make different step response, so an increasing amount of literature is devoted to proper tuning of PID controllers. The problem merits further investigation as traditional tuning methods make large control signal that can damages the system but using evolutionary algorithms based tuning methods improve the control signal and closed loop performance. In this paper three tuning methods for PID controllers have been studied namely Ziegler and Nichols, which is traditional tuning method and evolutionary algorithms based tuning methods, that are, Genetic algorithm and particle swarm optimization. To examine the validity of PSO and GA tuning methods a comparative analysis of DC motor plant is studied. Simulation results reveal that evolutionary algorithms based tuning method have improved control signal amplitude and quality factors of the closed loop system such as rise time, integral absolute error (IAE) and maximum overshoot.Keywords: evolutionary algorithm, genetic algorithm, particle swarm optimization, PID controller
Procedia PDF Downloads 483949 Sporting Events among the Disabled between Excellence and Ideal in Motor Performance: Analytical Descriptive Study in Some Paralympic Sports
Authors: Guebli Abdelkader, Reguieg Madani, Belkadi Adel, Sbaa Bouabdellah
Abstract:
The identification of mechanical variables in the motor performance trajectory has a prominent role in improving skill performance, error-exceeding, it contributes seriously to solving some problems of learning and training. The study aims to highlight the indicators of motor performance for Paralympic athletes during the practicing sports between modelling and between excellence in motor performance, this by taking into account the distinction of athlete practicing with special behavioral skills for the Paralympic athletes. In the study, we relied on the analysis of some previous research of biomechanical performance indicators during some of the events sports (shooting activities in the Paralympic athletics, shooting skill in the wheelchair basketball). The results of the study highlight the distinction of disabled practitioners of sporting events identified in motor performance during practice, by overcoming some physics indicators in human movement, as a lower center of body weight, increase in offset distance, such resistance which requires them to redouble their efforts. However, the results of the study highlighted the strength of the correlation between biomechanical variables of motor performance and the digital level achievement similar to the other practitioners normal.Keywords: sports, the disabled, motor performance, Paralympic
Procedia PDF Downloads 283948 Gluability of Bambusa balcooa and Bambusa vulgaris for Development of Laminated Panels
Authors: Daisy Biswas, Samar Kanti Bose, M. Mozaffar Hossain
Abstract:
The development of value added composite products from bamboo with the application of gluing technology can play a vital role in economic development and also in forest resource conservation of any country. In this study, the gluability of Bambusa balcooa and Bambusa vulgaris, two locally grown bamboo species of Bangladesh was assessed. As the culm wall thickness of bamboos decreases from bottom to top, a culm portion of up to 5.4 m and 3.6 m were used from the base of B. balcooa and B. vulgaris, respectively, to get rectangular strips of uniform thickness. The color of the B. vulgaris strips was yellowish brown and that of B. balcooa was reddish brown. The strips were treated in borax-boric, bleaching and carbonization for extending the service life of the laminates. The preservative treatments changed the color of the strips. Borax–boric acid treated strips were reddish brown. When bleached with hydrogen peroxide, the color of the strips turned into whitish yellow. Carbonization produced dark brownish strips having coffee flavor. Chemical constituents for untreated and treated strips were determined. B. vulgaris was more acidic than B. balcooa. Then the treated strips were used to develop three-layered bamboo laminated panel. Urea formaldehyde (UF) and polyvinyl acetate (PVA) were used as binder. The shear strength and abrasive resistance of the panel were evaluated. It was found that the shear strength of the UF-panel was higher than the PVA-panel for all treatments. Between the species, gluability of B. vulgaris was better and in some cases better than hardwood species. The abrasive resistance of B. balcooa is slightly higher than B. vulgaris; however, the latter was preferred as it showed well gluability. The panels could be used as structural panel, floor tiles, flat pack furniture component, and wall panel etc. However, further research on durability and creep behavior of the product in service condition is warranted.Keywords: Bambusa balcooa, Bambusa vulgaris, polyvinyl acetate, urea formaldehyde
Procedia PDF Downloads 262947 Quality Characteristics of Treated Wastewater of 'Industrial Area Foggia'
Authors: Grazia Disciglio, Annalisa Tarantino, Emanuele Tarantino
Abstract:
The production system of Foggia province (Apulia, Southern Italy) is characterized by the presence of numerous agro-food industries whose activities include the processing of vegetables products that release large quantities of wastewater. The reuse in agriculture of these wastewaters offers the opportunity to reduce the costs of their disposal and minimizing their environmental impact. In addition, in this area, which suffers from water shortage, the use of agro-industrial wastewater is essential in the very intensive irrigation cropping systems. The present investigation was carried out in years 2009 and 2010 to monitor the physico-chemical and microbiological characteristics of the industrial wastewater (IWW) from the secondary treatment plant of the 'Industrial Area of Foggia'. The treatment plant released on average about 567,000 m3y-1 of IWW, which distribution was not uniform over the year. The monthly values were about 250,000 m3 from November to June and about 90,000 m3 from July to October. The obtained results revealed that IWW was characterized by low values of Total Suspended Solids (TSS), Biological Oxygen Demand (BOD), Chemical Oxygen Demand (COD), Electrical Conductivity (EC) and Sodium Absorption Rate (SAR). An occasional presence of heavy metal and high concentration of total phosphorus, total nitrogen, ammoniacal nitrogen and microbial organisms (Escherichia coli and Salmonella) were observed. Due to the presence of this pathogenic microorganisms and sometimes of heavy metals, which may raise sanitary and environmental problems in order to the possible irrigation reuse of this IWW, a tertiary treatment of wastewater based on filtration and disinfection in line are recommended. Researches on the reuse of treated IWW on crops (olive, artichoke, industrial tomatoes, fennel, lettuce etc.) did not show significant differences among the irrigated plots for most of the soil and yield characteristics.Keywords: agroindustrial wastewater, irrigation, microbiological characteristic, physico-chemical characteristics
Procedia PDF Downloads 316946 System of Quality Automation for Documents (SQAD)
Authors: R. Babi Saraswathi, K. Divya, A. Habeebur Rahman, D. B. Hari Prakash, S. Jayanth, T. Kumar, N. Vijayarangan
Abstract:
Document automation is the design of systems and workflows, assembling repetitive documents to meet the specific business needs. In any organization or institution, documenting employee’s information is very important for both employees as well as management. It shows an individual’s progress to the management. Many documents of the employee are in the form of papers, so it is very difficult to arrange and for future reference we need to spend more time in getting the exact document. Also, it is very tedious to generate reports according to our needs. The process gets even more difficult on getting approvals and hence lacks its security aspects. This project overcomes the above-stated issues. By storing the details in the database and maintaining the e-documents, the automation system reduces the manual work to a large extent. Then the approval process of some important documents can be done in a much-secured manner by using Digital Signature and encryption techniques. Details are maintained in the database and e-documents are stored in specific folders and generation of various kinds of reports is possible. Moreover, an efficient search method is implemented is used in the database. Automation supporting document maintenance in many aspects is useful for minimize data entry, reduce the time spent on proof-reading, avoids duplication, and reduce the risks associated with the manual error, etc.Keywords: e-documents, automation, digital signature, encryption
Procedia PDF Downloads 391945 Random Subspace Neural Classifier for Meteor Recognition in the Night Sky
Authors: Carlos Vera, Tetyana Baydyk, Ernst Kussul, Graciela Velasco, Miguel Aparicio
Abstract:
This article describes the Random Subspace Neural Classifier (RSC) for the recognition of meteors in the night sky. We used images of meteors entering the atmosphere at night between 8:00 p.m.-5: 00 a.m. The objective of this project is to classify meteor and star images (with stars as the image background). The monitoring of the sky and the classification of meteors are made for future applications by scientists. The image database was collected from different websites. We worked with RGB-type images with dimensions of 220x220 pixels stored in the BitMap Protocol (BMP) format. Subsequent window scanning and processing were carried out for each image. The scan window where the characteristics were extracted had the size of 20x20 pixels with a scanning step size of 10 pixels. Brightness, contrast and contour orientation histograms were used as inputs for the RSC. The RSC worked with two classes and classified into: 1) with meteors and 2) without meteors. Different tests were carried out by varying the number of training cycles and the number of images for training and recognition. The percentage error for the neural classifier was calculated. The results show a good RSC classifier response with 89% correct recognition. The results of these experiments are presented and discussed.Keywords: contour orientation histogram, meteors, night sky, RSC neural classifier, stars
Procedia PDF Downloads 138944 Improvement in Blast Furnace Performance Using Softening - Melting Zone Profile Prediction Model at G Blast Furnace, Tata Steel Jamshedpur
Authors: Shoumodip Roy, Ankit Singhania, K. R. K. Rao, Ravi Shankar, M. K. Agarwal, R. V. Ramna, Uttam Singh
Abstract:
The productivity of a blast furnace and the quality of the hot metal produced are significantly dependent on the smoothness and stability of furnace operation. The permeability of the furnace bed, as well as the gas flow pattern, influences the steady control of process parameters. The softening – melting zone that is formed inside the furnace contributes largely in distribution of the gas flow and the bed permeability. A better shape of softening-melting zone enhances the performance of blast furnace, thereby reducing the fuel rates and improving furnace life. Therefore, predictive model of the softening- melting zone profile can be utilized to control and improve the furnace operation. The shape of softening-melting zone depends upon the physical and chemical properties of the agglomerates and iron ore charged in the furnace. The variations in the agglomerate proportion in the burden at G Blast furnace disturbed the furnace stability. During such circumstances, it was analyzed that a w-shape softening-melting zone profile was formed inside the furnace. The formation of w-shape zone resulted in poor bed permeability and non-uniform gas flow. There was a significant increase in the heat loss at the lower zone of the furnace. The fuel demand increased, and the huge production loss was incurred. Therefore, visibility of softening-melting zone profile was necessary in order to pro-actively optimize the process parameters and thereby to operate the furnace smoothly. Using stave temperatures, a model was developed that predicted the shape of the softening-melting zone inside the furnace. It was observed that furnace operated smoothly during inverse V-shape of the zone and vice-versa during w-shape. This model helped to control the heat loss, optimize the burden distribution and lower the fuel rate at G Blast Furnace, TSL Jamshedpur. As a result of furnace stabilization productivity increased by 10% and fuel rate reduced by 80 kg/thm. Details of the process have been discussed in this paper.Keywords: agglomerate, blast furnace, permeability, softening-melting
Procedia PDF Downloads 252943 A Model for Diagnosis and Prediction of Coronavirus Using Neural Network
Authors: Sajjad Baghernezhad
Abstract:
Meta-heuristic and hybrid algorithms have high adeer in modeling medical problems. In this study, a neural network was used to predict covid-19 among high-risk and low-risk patients. This study was conducted to collect the applied method and its target population consisting of 550 high-risk and low-risk patients from the Kerman University of medical sciences medical center to predict the coronavirus. In this study, the memetic algorithm, which is a combination of a genetic algorithm and a local search algorithm, has been used to update the weights of the neural network and develop the accuracy of the neural network. The initial study showed that the accuracy of the neural network was 88%. After updating the weights, the memetic algorithm increased by 93%. For the proposed model, sensitivity, specificity, positive predictivity value, value/accuracy to 97.4, 92.3, 95.8, 96.2, and 0.918, respectively; for the genetic algorithm model, 87.05, 9.20 7, 89.45, 97.30 and 0.967 and for logistic regression model were 87.40, 95.20, 93.79, 0.87 and 0.916. Based on the findings of this study, neural network models have a lower error rate in the diagnosis of patients based on individual variables and vital signs compared to the regression model. The findings of this study can help planners and health care providers in signing programs and early diagnosis of COVID-19 or Corona.Keywords: COVID-19, decision support technique, neural network, genetic algorithm, memetic algorithm
Procedia PDF Downloads 66942 Investigation of Dry Ice Mixed Novel Hybrid Lubri-Coolant in Sustainable Machining of Ti-6AL-4V Alloy: A Comparison of Experimental and Modelling
Authors: Muhammad Jamil, Ning He, Aqib Mashood Khan, Munish Kumar Gupta
Abstract:
Ti-6Al-4V has numerous applications in the medical, automobile, and aerospace industries due to corrosion resistivity, structural stability, and chemical inertness to most fluids at room temperature. These peculiar characteristics are beneficial for their application and present formidable challenges during machining. Machining of Ti-6Al-4V produces an elevated cutting temperature above 1000oC at dry conditions. This accelerates tool wear and reduces product quality. Therefore, there is always a need to employ sustainable/effective coolant/lubricant when machining such alloy. In this study, Finite Element Modeling (FEM) and experimental analysis when cutting Ti-6Al-4V under a distinctly developed dry ice mixed hybrid lubri-coolant are presented. This study aims to model the milling process of Ti-6Al-4V under a proposed novel hybrid lubri-coolant using different cutting speeds and feed per tooth DEFORM® software package was used to conduct the FEM and the numerical model was experimentally validated. A comparison of experimental and simulation results showed a maximum error of no more than 6% for all experimental conditions. In a nutshell, it can be said that the proposed model is effective in predicting the machining temperature precisely.Keywords: friction coefficient, heat transfer, finite element modeling (FEM), milling Ti-6Al-4V
Procedia PDF Downloads 58941 Transient Enhanced LDO Voltage Regulator with Improved Feed Forward Path Compensation
Authors: A. Suresh, Sreehari Rao Patri, K. S. R. Krishnaprasad
Abstract:
An ultra low power capacitor less low-dropout voltage regulator with improved transient response using gain enhanced feed forward path compensation is presented in this paper. It is based on a cascade of a voltage amplifier and a transconductor stage in the feed forward path with regular error amplifier to form a composite gain-enhanced feed forward stage. It broadens the gain bandwidth and thus improves the transient response without substantial increase in power consumption. The proposed LDO, designed for a maximum output current of 100 mA in UMC 180 nm, requires a quiescent current of 69 µA. An undershoot of 153.79mV for a load current changes from 0mA to 100mA and an overshoot of 196.24mV for current change of 100mA to 0mA. The settling time is approximately 1.1 µs for the output voltage undershoot case. The load regulation is of 2.77 µV/mA at load current of 100mA. Reference voltage is generated by using an accurate band gap reference circuit of 0.8V.The costly features of SOC such as total chip area and power consumption is drastically reduced by the use of only a total compensation capacitance of 6pF while consuming power consumption of 0.096 mW.Keywords: capacitor-less LDO, frequency compensation, transient response, latch, self-biased differential amplifier
Procedia PDF Downloads 451940 Efficiency of Google Translate and Bing Translator in Translating Persian-to-English Texts
Authors: Samad Sajjadi
Abstract:
Machine translation is a new subject increasingly being used by academic writers, especially students and researchers whose native language is not English. There are numerous studies conducted on machine translation, but few investigations have assessed the accuracy of machine translation from Persian to English at lexical, semantic, and syntactic levels. Using Groves and Mundt’s (2015) Model of error taxonomy, the current study evaluated Persian-to-English translations produced by two famous online translators, Google Translate and Bing Translator. A total of 240 texts were randomly selected from different academic fields (law, literature, medicine, and mass media), and 60 texts were considered for each domain. All texts were rendered by the two translation systems and then by four human translators. All statistical analyses were applied using SPSS. The results indicated that Google translations were more accurate than the translations produced by the Bing Translator, especially in the domains of medicine (lexis: 186 vs. 225; semantic: 44 vs. 48; syntactic: 148 vs. 264 errors) and mass media (lexis: 118 vs. 149; semantic: 25 vs. 32; syntactic: 110 vs. 220 errors), respectively. Nonetheless, both machines are reasonably accurate in Persian-to-English translation of lexicons and syntactic structures, particularly from mass media and medical texts.Keywords: machine translations, accuracy, human translation, efficiency
Procedia PDF Downloads 77939 Enhancing Academic Writing Through Artificial Intelligence: Opportunities and Challenges
Authors: Abubakar Abdulkareem, Nasir Haruna Soba
Abstract:
Artificial intelligence (AI) is developing at a rapid pace, revolutionizing several industries, including education. This talk looks at how useful AI can be for academic writing, with an emphasis on how it can help researchers be more accurate, productive, and creative. The academic world now relies heavily on AI technologies like grammar checkers, plagiarism detectors, and content generators to help with the writing, editing, and formatting of scholarly papers. This study explores the particular uses of AI in academic writing and assesses how useful and helpful these applications may be for both students and scholars. By means of an extensive examination of extant literature and a sequence of empirical case studies, we scrutinize the merits and demerits of artificial intelligence tools utilized in academic writing. Important discoveries indicate that although AI greatly increases productivity and lowers human error, there are still issues that need to be resolved, including reliance, ethical concerns, and the potential loss of critical thinking abilities. The talk ends with suggestions for incorporating AI tools into academic settings so that they enhance rather than take the place of the intellectual rigor that characterizes scholarly work. This study adds to the continuing conversation about artificial intelligence (AI) in higher education by supporting a methodical strategy that uses technology to enhance human abilities in academic writing.Keywords: artificial intelligence, academic writing, ai tools, productivity, ethics, higher education
Procedia PDF Downloads 27938 Comparison of Sensitivity and Specificity of Pap Smear and Polymerase Chain Reaction Methods for Detection of Human Papillomavirus: A Review of Literature
Authors: M. Malekian, M. E. Heydari, M. Irani Estyar
Abstract:
Human papillomavirus (HPV) is one of the most common sexually transmitted infection, which may lead to cervical cancer as the main cause of it. With early diagnosis and treatment in health care services, cervical cancer and its complications are considered to be preventable. This study was aimed to compare the efficiency, sensitivity, and specificity of Pap smear and polymerase chain reaction (PCR) in detecting HPV. A literature search was performed in Google Scholar, PubMed and SID databases using the keywords 'human papillomavirus', 'pap smear' and 'polymerase change reaction' to identify studies comparing Pap smear and PCR methods for the detection. No restrictions were considered.10 studies were included in this review. All samples that were positive by pop smear were also positive by PCR. However, there were positive samples detected by PCR which was negative by pop smear and in all studies, many positive samples were missed by pop smear technique. Although The Pap smear had high specificity, PCR based HPV detection was more sensitive method and had the highest sensitivity. In order to promote the quality of detection and high achievement of the maximum results, PCR diagnostic methods in addition to the Pap smear are needed and Pap smear method should be combined with PCR techniques according to the high error rate of Pap smear in detection.Keywords: human papillomavirus, cervical cancer, pap smear, polymerase chain reaction
Procedia PDF Downloads 131937 Triangular Geometric Feature for Offline Signature Verification
Authors: Zuraidasahana Zulkarnain, Mohd Shafry Mohd Rahim, Nor Anita Fairos Ismail, Mohd Azhar M. Arsad
Abstract:
Handwritten signature is accepted widely as a biometric characteristic for personal authentication. The use of appropriate features plays an important role in determining accuracy of signature verification; therefore, this paper presents a feature based on the geometrical concept. To achieve the aim, triangle attributes are exploited to design a new feature since the triangle possesses orientation, angle and transformation that would improve accuracy. The proposed feature uses triangulation geometric set comprising of sides, angles and perimeter of a triangle which is derived from the center of gravity of a signature image. For classification purpose, Euclidean classifier along with Voting-based classifier is used to verify the tendency of forgery signature. This classification process is experimented using triangular geometric feature and selected global features. Based on an experiment that was validated using Grupo de Senales 960 (GPDS-960) signature database, the proposed triangular geometric feature achieves a lower Average Error Rates (AER) value with a percentage of 34% as compared to 43% of the selected global feature. As a conclusion, the proposed triangular geometric feature proves to be a more reliable feature for accurate signature verification.Keywords: biometrics, euclidean classifier, features extraction, offline signature verification, voting-based classifier
Procedia PDF Downloads 378936 Bidirectional Long Short-Term Memory-Based Signal Detection for Orthogonal Frequency Division Multiplexing With All Index Modulation
Authors: Mahmut Yildirim
Abstract:
This paper proposed the bidirectional long short-term memory (Bi-LSTM) network-aided deep learning (DL)-based signal detection for Orthogonal frequency division multiplexing with all index modulation (OFDM-AIM), namely Bi-DeepAIM. OFDM-AIM is developed to increase the spectral efficiency of OFDM with index modulation (OFDM-IM), a promising multi-carrier technique for communication systems beyond 5G. In this paper, due to its strong classification ability, Bi-LSTM is considered an alternative to the maximum likelihood (ML) algorithm, which is used for signal detection in the classical OFDM-AIM scheme. The performance of the Bi-DeepAIM is compared with LSTM network-aided DL-based OFDM-AIM (DeepAIM) and classic OFDM-AIM that uses (ML)-based signal detection via BER performance and computational time criteria. Simulation results show that Bi-DeepAIM obtains better bit error rate (BER) performance than DeepAIM and lower computation time in signal detection than ML-AIM.Keywords: bidirectional long short-term memory, deep learning, maximum likelihood, OFDM with all index modulation, signal detection
Procedia PDF Downloads 72935 Data-Driven Approach to Predict Inpatient's Estimated Discharge Date
Authors: Ayliana Dharmawan, Heng Yong Sheng, Zhang Xiaojin, Tan Thai Lian
Abstract:
To facilitate discharge planning, doctors are presently required to assign an Estimated Discharge Date (EDD) for each patient admitted to the hospital. This assignment of the EDD is largely based on the doctor’s judgment. This can be difficult for cases which are complex or relatively new to the doctor. It is hypothesized that a data-driven approach would be able to facilitate the doctors to make accurate estimations of the discharge date. Making use of routinely collected data on inpatient discharges between January 2013 and May 2016, a predictive model was developed using machine learning techniques to predict the Length of Stay (and hence the EDD) of inpatients, at the point of admission. The predictive performance of the model was compared to that of the clinicians using accuracy measures. Overall, the best performing model was found to be able to predict EDD with an accuracy improvement in Average Squared Error (ASE) by -38% as compared to the first EDD determined by the present method. It was found that important predictors of the EDD include the provisional diagnosis code, patient’s age, attending doctor at admission, medical specialty at admission, accommodation type, and the mean length of stay of the patient in the past year. The predictive model can be used as a tool to accurately predict the EDD.Keywords: inpatient, estimated discharge date, EDD, prediction, data-driven
Procedia PDF Downloads 174934 An Automatic Generating Unified Modelling Language Use Case Diagram and Test Cases Based on Classification Tree Method
Authors: Wassana Naiyapo, Atichat Sangtong
Abstract:
The processes in software development by Object Oriented methodology have many stages those take time and high cost. The inconceivable error in system analysis process will affect to the design and the implementation process. The unexpected output causes the reason why we need to revise the previous process. The more rollback of each process takes more expense and delayed time. Therefore, the good test process from the early phase, the implemented software is efficient, reliable and also meet the user’s requirement. Unified Modelling Language (UML) is the tool which uses symbols to describe the work process in Object Oriented Analysis (OOA). This paper presents the approach for automatically generated UML use case diagram and test cases. UML use case diagram is generated from the event table and test cases are generated from use case specifications and Graphic User Interfaces (GUI). Test cases are derived from the Classification Tree Method (CTM) that classify data to a node present in the hierarchy structure. Moreover, this paper refers to the program that generates use case diagram and test cases. As the result, it can reduce work time and increase efficiency work.Keywords: classification tree method, test case, UML use case diagram, use case specification
Procedia PDF Downloads 162933 Bioinformatics Approach to Identify Physicochemical and Structural Properties Associated with Successful Cell-free Protein Synthesis
Authors: Alexander A. Tokmakov
Abstract:
Cell-free protein synthesis is widely used to synthesize recombinant proteins. It allows genome-scale expression of various polypeptides under strictly controlled uniform conditions. However, only a minor fraction of all proteins can be successfully expressed in the systems of protein synthesis that are currently used. The factors determining expression success are poorly understood. At present, the vast volume of data is accumulated in cell-free expression databases. It makes possible comprehensive bioinformatics analysis and identification of multiple features associated with successful cell-free expression. Here, we describe an approach aimed at identification of multiple physicochemical and structural properties of amino acid sequences associated with protein solubility and aggregation and highlight major correlations obtained using this approach. The developed method includes: categorical assessment of the protein expression data, calculation and prediction of multiple properties of expressed amino acid sequences, correlation of the individual properties with the expression scores, and evaluation of statistical significance of the observed correlations. Using this approach, we revealed a number of statistically significant correlations between calculated and predicted features of protein sequences and their amenability to cell-free expression. It was found that some of the features, such as protein pI, hydrophobicity, presence of signal sequences, etc., are mostly related to protein solubility, whereas the others, such as protein length, number of disulfide bonds, content of secondary structure, etc., affect mainly the expression propensity. We also demonstrated that amenability of polypeptide sequences to cell-free expression correlates with the presence of multiple sites of post-translational modifications. The correlations revealed in this study provide a plethora of important insights into protein folding and rationalization of protein production. The developed bioinformatics approach can be of practical use for predicting expression success and optimizing cell-free protein synthesis.Keywords: bioinformatics analysis, cell-free protein synthesis, expression success, optimization, recombinant proteins
Procedia PDF Downloads 419932 Ambiguity Resolution for Ground-based Pulse Doppler Radars Using Multiple Medium Pulse Repetition Frequency
Authors: Khue Nguyen Dinh, Loi Nguyen Van, Thanh Nguyen Nhu
Abstract:
In this paper, we propose an adaptive method to resolve ambiguities and a ghost target removal process to extract targets detected by a ground-based pulse-Doppler radar using medium pulse repetition frequency (PRF) waveforms. The ambiguity resolution method is an adaptive implementation of the coincidence algorithm, which is implemented on a two-dimensional (2D) range-velocity matrix to resolve range and velocity ambiguities simultaneously, with a proposed clustering filter to enhance the anti-error ability of the system. Here we consider the scenario of multiple target environments. The ghost target removal process, which is based on the power after Doppler processing, is proposed to mitigate ghosting detections to enhance the performance of ground-based radars using a short PRF schedule in multiple target environments. Simulation results on a ground-based pulsed Doppler radar model will be presented to show the effectiveness of the proposed approach.Keywords: ambiguity resolution, coincidence algorithm, medium PRF, ghosting removal
Procedia PDF Downloads 151931 Encryption and Decryption of Nucleic Acid Using Deoxyribonucleic Acid Algorithm
Authors: Iftikhar A. Tayubi, Aabdulrahman Alsubhi, Abdullah Althrwi
Abstract:
The deoxyribonucleic acid text provides a single source of high-quality Cryptography about Deoxyribonucleic acid sequence for structural biologists. We will provide an intuitive, well-organized and user-friendly web interface that allows users to encrypt and decrypt Deoxy Ribonucleic Acid sequence text. It includes complex, securing by using Algorithm to encrypt and decrypt Deoxy Ribonucleic Acid sequence. The utility of this Deoxy Ribonucleic Acid Sequence Text is that, it can provide a user-friendly interface for users to Encrypt and Decrypt store the information about Deoxy Ribonucleic Acid sequence. These interfaces created in this project will satisfy the demands of the scientific community by providing fully encrypt of Deoxy Ribonucleic Acid sequence during this website. We have adopted a methodology by using C# and Active Server Page.NET for programming which is smart and secure. Deoxy Ribonucleic Acid sequence text is a wonderful piece of equipment for encrypting large quantities of data, efficiently. The users can thus navigate from one encoding and store orange text, depending on the field for user’s interest. Algorithm classification allows a user to Protect the deoxy ribonucleic acid sequence from change, whether an alteration or error occurred during the Deoxy Ribonucleic Acid sequence data transfer. It will check the integrity of the Deoxy Ribonucleic Acid sequence data during the access.Keywords: algorithm, ASP.NET, DNA, encrypt, decrypt
Procedia PDF Downloads 234930 Legal Aspects in Character Merchandising with Reference to Right to Image of Celebrities
Authors: W. R. M. Shehani Shanika
Abstract:
Selling goods and services using images, names and personalities of celebrities has become a common marketing strategy identified in modern physical and online markets. Two concepts called globalization and open economy have given numerous reasons to develop businesses to earn higher profits. Therefore, global market plus domestic markets in various countries have vigorously endorsing images of famous sport stars, film stars, singing stars and cartoon characters for the purpose of increasing demand for goods and services rendered by them. It has been evident that these trade strategies have become a threat to famous personalities in financially and personally. Right to the image is a basic human right which celebrities owned to avoid themselves from various commercial exploitations. In this respect, this paper aims to assess whether the law relating to character merchandising satisfactorily protects right to image of celebrities. However, celebrities can decide how much they receive for each representation to the general public. Simply they have exclusive right to decide monetary value for their image. But most commonly every country uses law relating to unfair competition to regulate matters arise thereof. Legal norms in unfair competition are not enough to protect image of celebrities. Therefore, celebrities must be able to avoid unauthorized use of their images for commercial purposes by fraudulent traders and getting unjustly enriched, as their images have economic value. They have the right for use their image for any commercial purpose and earn profits. Therefore it is high time to recognize right to image as a new dimension to be protected in the legal framework of character merchandising. Unfortunately, to the author’s best knowledge there are no any uniform, single international standard which recognizes right to the image of celebrities in the context of character merchandising. The paper identifies it as a controversial legal barrier faced by celebrities in the rapidly evolving marketplace. Finally, this library-based research concludes with proposals to ensure the right to image more broadly in the legal context of character merchandising.Keywords: brand endorsement, celebrity, character merchandising, intellectual property rights, right to image, unfair competition
Procedia PDF Downloads 138929 Vertical Accuracy Evaluation of Indian National DEM (CartoDEM v3) Using Dual Frequency GNSS Derived Ground Control Points for Lower Tapi Basin, Western India
Authors: Jaypalsinh B. Parmar, Pintu Nakrani, Ashish Chaurasia
Abstract:
Digital Elevation Model (DEM) is considered as an important data in GIS-based terrain analysis for many applications and assessment of processes such as environmental and climate change studies, hydrologic modelling, etc. Vertical accuracy of DEM having geographically dynamic nature depends on different parameters which affect the model simulation outcomes. Vertical accuracy assessment in Indian landscape especially in low-lying coastal urban terrain such as lower Tapi Basin is very limited. In the present study, attempt has been made to evaluate the vertical accuracy of 30m resolution open source Indian National Cartosat-1 DEM v3 for Lower Tapi Basin (LTB) from western India. The extensive field investigation is carried out using stratified random fast static DGPS survey in the entire study region, and 117 high accuracy ground control points (GCPs) have been obtained. The above open source DEM was compared with obtained GCPs, and different statistical attributes were envisaged, and vertical error histograms were also evaluated.Keywords: CartoDEM, Digital Elevation Model, GPS, lower Tapi basin
Procedia PDF Downloads 358928 Morphology and Risk Factors for Blunt Aortic Trauma in Car Accidents: An Autopsy Study
Authors: Ticijana Prijon, Branko Ermenc
Abstract:
Background: Blunt aortic trauma (BAT) includes various morphological changes that occur during deceleration, acceleration and/or body compression in traffic accidents. The various forms of BAT, from limited laceration of the intima to complete transection of the aorta, depends on the force acting on the vessel wall and the tolerance of the aorta to injury. The force depends on the change in velocity, the dynamics of the accident and of the seating position in the car. Tolerance to aortic injury depends on the anatomy, histological structure and pathomorphological alterations due to aging or disease of the aortic wall.An overview of the literature and medical documentation reveals that different terms are used to describe certain forms of BAT, which can lead to misinterpretation of findings or diagnoses. We therefore, propose a classification that would enable uniform systematic screening of all forms of BAT. We have classified BAT into three morphologycal types: TYPE I (intramural), TYPE II (transmural) and TYPE III (multiple) aortic ruptures with appropriate subtypes. Methods: All car accident casualties examined at the Institute of Forensic Medicine from 2001 to 2009 were included in this retrospective study. Autopsy reports were used to determine the occurrence of each morphological type of BAT in deceased drivers, front seat passengers and other passengers in cars and to define the morphology of BAT in relation to the accident dynamics and the age of the fatalities. Results: A total of 391 fatalities in car accidents were included in the study. TYPE I, TYPE II and TYPE III BAT were observed in 10,9%, 55,6% and 33,5%, respectively. The incidence of BAT in drivers, front seat and other passengers was 36,7%, 43,1% and 28,6%, respectively. In frontal collisions, the incidence of BAT was 32,7%, in lateral collisions 54,2%, and in other traffic accidents 29,3%. The average age of fatalities with BAT was 42,8 years and of those without BAT 39,1 years. Conclusion: Identification and early recognition of the risk factors of BAT following a traffic accident is crucial for successful treatment of patients with BAT. Front seat passengers over 50 years of age who have been injured in a lateral collision are the most at risk of BAT.Keywords: aorta, blunt trauma, car accidents, morphology, risk factors
Procedia PDF Downloads 513927 The Effect of Nursing Teamwork Training on Nursing Teamwork Effectiveness
Authors: Manar Ahmed Elbadawy
Abstract:
Background: Empirical evidence suggested that improving nursing teamwork (NTW) may be the key to reducing medical error. The functioning nursing teams require open communication, mutual respect, and shared mental models to activate quality patient care. The complexity and the high demands for specialized nursing knowledge and skill also require nursing staff to consult with one another and work in teams regularly. The current study aimed to evaluate the effect of the nursing teamwork training program on nursing teamwork effectiveness. Design: A quasi-experimental (one group pretest-posttest) design was utilized. Three medical intensive care units at a teaching hospital affiliated to Cairo University Hospital, Egypt. Subjects: A convenient sample of 48 nursing staff worked at the selected units. The Nursing Teamwork Observational Checklist was used. Results: Total (NTW) mean scores exhibited quite elevation post-program implementation compared to preprogram and showed little decrease 3 months later ( = 2.52, SD = ± 0.27, mean % =51.98, = 2.72, SD = ± 0.20, mean %=72.45, = 2.67, SD = ± 0.11, mean %= 67.48 respectively). Conclusion: Implementation of (NTW) training program had a positive effect on increasing (NTW) effectiveness. Regular and frequent short-term teamwork training is important to be introduced as well as sustainable monitoring is required to ensure nursing attitudes, knowledge and skills’ change about teamwork effectiveness.Keywords: effectiveness, nursing, teamwork, training
Procedia PDF Downloads 123