Search results for: automatic classification
572 Modeling Loads Applied to Main and Crank Bearings in the Compression-Ignition Two-Stroke Engine
Authors: Marcin Szlachetka, Mateusz Paszko, Grzegorz Baranski
Abstract:
This paper discusses the AVL EXCITE Designer simulation research into loads applied to main and crank bearings in the compression-ignition two-stroke engine. There was created a model of engine lubrication system which covers the part of this system related to particular nodes of a bearing system, i.e. a connection of main bearings in an engine block with a crankshaft, a connection of crank pins with a connecting rod. The analysis focused on the load given as a distribution of hydrodynamic oil film pressure corresponding different values of radial internal clearance. There was also studied the impact of gas force on minimal oil film thickness in main and crank bearings versus crankshaft rotational speed. Our model calculates oil film parameters, an oil film pressure distribution, an oil temperature change and dimensions of bearings as well as an oil temperature distribution on surfaces of bearing seats. Accordingly, it was possible to select, for example, a correct clearance for each of the node bearings. The research was performed for several values of engine crankshaft speed ranging from 800 RPM to 4000 RPM. Bearing oil pressure was changed according to engine speed ranging between 1 bar and 5 bar and an oil temperature of 90°C. The main bearing clearances made initially for the calculation and research were: 0.015 mm, 0.025 mm, 0.035 mm, 0.05 mm, 0.1 mm. The oil used for the research corresponded the SAE 5W-40 classification. The paper presents the selected research results referring to certain specific operating points and bearing radial internal clearances. Acknowledgement: This work has been realized in the cooperation with The Construction Office of WSK ‘PZL-KALISZ’ S.A. and is part of Grant Agreement No. POIR.01.02.00-00-0002/15 financed by the Polish National Centre for Research and Development.Keywords: crank bearings, diesel engine, oil film, two-stroke engine
Procedia PDF Downloads 214571 A Geometric Based Hybrid Approach for Facial Feature Localization
Authors: Priya Saha, Sourav Dey Roy Jr., Debotosh Bhattacharjee, Mita Nasipuri, Barin Kumar De, Mrinal Kanti Bhowmik
Abstract:
Biometric face recognition technology (FRT) has gained a lot of attention due to its extensive variety of applications in both security and non-security perspectives. It has come into view to provide a secure solution in identification and verification of person identity. Although other biometric based methods like fingerprint scans, iris scans are available, FRT is verified as an efficient technology for its user-friendliness and contact freeness. Accurate facial feature localization plays an important role for many facial analysis applications including biometrics and emotion recognition. But, there are certain factors, which make facial feature localization a challenging task. On human face, expressions can be seen from the subtle movements of facial muscles and influenced by internal emotional states. These non-rigid facial movements cause noticeable alterations in locations of facial landmarks, their usual shapes, which sometimes create occlusions in facial feature areas making face recognition as a difficult problem. The paper proposes a new hybrid based technique for automatic landmark detection in both neutral and expressive frontal and near frontal face images. The method uses the concept of thresholding, sequential searching and other image processing techniques for locating the landmark points on the face. Also, a Graphical User Interface (GUI) based software is designed that could automatically detect 16 landmark points around eyes, nose and mouth that are mostly affected by the changes in facial muscles. The proposed system has been tested on widely used JAFFE and Cohn Kanade database. Also, the system is tested on DeitY-TU face database which is created in the Biometrics Laboratory of Tripura University under the research project funded by Department of Electronics & Information Technology, Govt. of India. The performance of the proposed method has been done in terms of error measure and accuracy. The method has detection rate of 98.82% on JAFFE database, 91.27% on Cohn Kanade database and 93.05% on DeitY-TU database. Also, we have done comparative study of our proposed method with other techniques developed by other researchers. This paper will put into focus emotion-oriented systems through AU detection in future based on the located features.Keywords: biometrics, face recognition, facial landmarks, image processing
Procedia PDF Downloads 412570 Development of Liquefaction-Induced Ground Damage Maps for the Wairau Plains, New Zealand
Authors: Omer Altaf, Liam Wotherspoon, Rolando Orense
Abstract:
The Wairau Plains are located in the north-east of the South Island of New Zealand in the region of Marlborough. The region is cut by many active crustal faults such as the Wairau, Awatere, and Clarence faults, which give rise to frequent seismic events. This paper presents the preliminary results of the overall project in which liquefaction-induced ground damage maps are developed in the Wairau Plains based on the Ministry of Business, Innovation and Employment NZ guidance. A suite of maps has been developed in relation to the level of details that was available to inform the liquefaction hazard mapping. Maps at the coarsest level of detail make use of regional geologic information, applying semi-quantitative criteria based on geological age, design peak ground accelerations and depth to the water table. The next level of detail incorporates higher resolution surface geomorphologic characteristics to better delineate potentially liquefiable and non-liquefiable deposits across the region. The most detailed assessment utilised CPT sounding data to develop ground damage response curves for areas across the region and provide a finer level of categorisation of liquefaction vulnerability. Linking these with design level earthquakes defined through NZGS guidelines will enable detailed classification to be carried out at CPT investigation locations, from very low through to high liquefaction vulnerability. To update classifications to these detailed levels, CPT investigations in geomorphic regions are grouped together to provide an indication of the representative performance of the soils in these areas making use of the geomorphic mapping outlined above.Keywords: hazard, liquefaction, mapping, seismicity
Procedia PDF Downloads 139569 A Conceptual Model of the 'Driver – Highly Automated Vehicle' System
Authors: V. A. Dubovsky, V. V. Savchenko, A. A. Baryskevich
Abstract:
The current trend in the automotive industry towards automatic vehicles is creating new challenges related to human factors. This occurs due to the fact that the driver is increasingly relieved of the need to be constantly involved in driving the vehicle, which can negatively impact his/her situation awareness when manual control is required, and decrease driving skills and abilities. These new problems need to be studied in order to provide road safety during the transition towards self-driving vehicles. For this purpose, it is important to develop an appropriate conceptual model of the interaction between the driver and the automated vehicle, which could serve as a theoretical basis for the development of mathematical and simulation models to explore different aspects of driver behaviour in different road situations. Well-known driver behaviour models describe the impact of different stages of the driver's cognitive process on driving performance but do not describe how the driver controls and adjusts his actions. A more complete description of the driver's cognitive process, including the evaluation of the results of his/her actions, will make it possible to more accurately model various aspects of the human factor in different road situations. This paper presents a conceptual model of the 'driver – highly automated vehicle' system based on the P.K. Anokhin's theory of functional systems, which is a theoretical framework for describing internal processes in purposeful living systems based on such notions as goal, desired and actual results of the purposeful activity. A central feature of the proposed model is a dynamic coupling mechanism between the decision-making of a driver to perform a particular action and changes of road conditions due to driver’s actions. This mechanism is based on the stage by stage evaluation of the deviations of the actual values of the driver’s action results parameters from the expected values. The overall functional structure of the highly automated vehicle in the proposed model includes a driver/vehicle/environment state analyzer to coordinate the interaction between driver and vehicle. The proposed conceptual model can be used as a framework to investigate different aspects of human factors in transitions between automated and manual driving for future improvements in driving safety, and for understanding how driver-vehicle interface must be designed for comfort and safety. A major finding of this study is the demonstration that the theory of functional systems is promising and has the potential to describe the interaction of the driver with the vehicle and the environment.Keywords: automated vehicle, driver behavior, human factors, human-machine system
Procedia PDF Downloads 145568 The Quantum Theory of Music and Human Languages
Authors: Mballa Abanda Luc Aurelien Serge, Henda Gnakate Biba, Kuate Guemo Romaric, Akono Rufine Nicole, Zabotom Yaya Fadel Biba, Petfiang Sidonie, Bella Suzane Jenifer
Abstract:
The main hypotheses proposed around the definition of the syllable and of music, of the common origin of music and language, should lead the reader to reflect on the cross-cutting questions raised by the debate on the notion of universals in linguistics and musicology. These are objects of controversy, and there lies its interest: the debate raises questions that are at the heart of theories on language. It is an inventive, original, and innovative research thesis. A contribution to the theoretical, musicological, ethno musicological, and linguistic conceptualization of languages, giving rise to the practice of interlocution between the social and cognitive sciences, the activities of artistic creation, and the question of modeling in the human sciences: mathematics, computer science, translation automation, and artificial intelligence. When you apply this theory to any text of a folksong of a world-tone language, you do not only piece together the exact melody, rhythm, and harmonies of that song as if you knew it in advance but also the exact speaking of this language. The author believes that the issue of the disappearance of tonal languages and their preservation has been structurally resolved, as well as one of the greatest cultural equations related to the composition and creation of tonal, polytonal, and random music. The experimentation confirming the theorization, I designed a semi-digital, semi-analog application that translates the tonal languages of Africa (about 2,100 languages) into blues, jazz, world music, polyphonic music, tonal and anatonal music, and deterministic and random music). To test this application, I use music reading and writing software that allows me to collect the data extracted from my mother tongue, which is already modeled in the musical staves saved in the ethnographic (semiotic) dictionary for automatic translation ( volume 2 of the book). The translation is done (from writing to writing, from writing to speech, and from writing to music). Mode of operation: you type a text on your computer, a structured song (chorus-verse), and you command the machine a melody of blues, jazz, and world music or variety, etc. The software runs, giving you the option to choose harmonies, and then you select your melody.Keywords: language, music, sciences, quantum entenglement
Procedia PDF Downloads 77567 A Study of Binding Methods and Techniques in Safavid Era Emphasizing on Iran Shahnamehs (16-18th Century AD/10-12th Century AH)
Authors: Ashrafosadat Mousavi Laer, Elaheh Moravej
Abstract:
The art of binding was simple and elementary at the beginning of Islam. This art thrived gradually and continued its development as an independent art. Identification of the binding techniques and used materials in covers and investigation of the arrays give us indexes for the better identification of different doctrines and methods of that time. The catalogers of the manuscripts usually pay attention to four items: gender, color, art elegances, injury, and exquisiteness of the cover. The criterion for classification of the covers is their art nature and gender. 15th century AD (9th century AH) was the period of the binding art development in which the most beautiful covers were produced by the so-called method of ‘burning’. At 16th century AD (10th century AH), in Safavid era, art changed completely and a fundamental evolution occurred in the technique and method of binding. The greatest change in this art was the extensive use of stamp that was made mostly of steel and copper. Theses stamps were presses against leather. These covers were called ‘beat’. In this paper, writing and bookbinding of about 32 Shahnamehs of Safavid era available in the Iranian libraries and museums are studied. An analytical-statistical study shows that four methods have been used including beat, burning, mosaic, and oily. 69 percent of the covers of these copies are cardboards with a leathery coating (goatskin) and have been produced by burning and beat methods. Its reasons are that these two methods have been common methods in Safavid era and performing them was only feasible on leather and the most desirable and commonly used leather of that time was goatskin which was the best option for cover legend durability and preserving the book and it was more durable because it had been made of goat skin. In addition, it had prepared a suitable opportunity for the binding artist’s creativity and innovation.Keywords: Shahnameh, Safavid era, bookbinding, beat cover, burning cover
Procedia PDF Downloads 238566 Using Visualization Techniques to Support Common Clinical Tasks in Clinical Documentation
Authors: Jonah Kenei, Elisha Opiyo
Abstract:
Electronic health records, as a repository of patient information, is nowadays the most commonly used technology to record, store and review patient clinical records and perform other clinical tasks. However, the accurate identification and retrieval of relevant information from clinical records is a difficult task due to the unstructured nature of clinical documents, characterized in particular by a lack of clear structure. Therefore, medical practice is facing a challenge thanks to the rapid growth of health information in electronic health records (EHRs), mostly in narrative text form. As a result, it's becoming important to effectively manage the growing amount of data for a single patient. As a result, there is currently a requirement to visualize electronic health records (EHRs) in a way that aids physicians in clinical tasks and medical decision-making. Leveraging text visualization techniques to unstructured clinical narrative texts is a new area of research that aims to provide better information extraction and retrieval to support clinical decision support in scenarios where data generated continues to grow. Clinical datasets in electronic health records (EHR) offer a lot of potential for training accurate statistical models to classify facets of information which can then be used to improve patient care and outcomes. However, in many clinical note datasets, the unstructured nature of clinical texts is a common problem. This paper examines the very issue of getting raw clinical texts and mapping them into meaningful structures that can support healthcare professionals utilizing narrative texts. Our work is the result of a collaborative design process that was aided by empirical data collected through formal usability testing.Keywords: classification, electronic health records, narrative texts, visualization
Procedia PDF Downloads 118565 Assessment of the Impacts of Climate Change on Climatic Zones over the Korean Peninsula for Natural Disaster Management Information
Authors: Sejin Jung, Dongho Kang, Byungsik Kim
Abstract:
Assessing the impact of climate change requires the use of a multi-model ensemble (MME) to quantify uncertainties between scenarios and produce downscaled outlines for simulation of climate under the influence of different factors, including topography. This study decreases climate change scenarios from the 13 global climate models (GCMs) to assess the impacts of future climate change. Unlike South Korea, North Korea lacks in studies using climate change scenarios of the CoupledModelIntercomparisonProject (CMIP5), and only recently did the country start the projection of extreme precipitation episodes. One of the main purposes of this study is to predict changes in the average climatic conditions of North Korea in the future. The result of comparing downscaled climate change scenarios with observation data for a reference period indicates high applicability of the Multi-Model Ensemble (MME). Furthermore, the study classifies climatic zones by applying the Köppen-Geiger climate classification system to the MME, which is validated for future precipitation and temperature. The result suggests that the continental climate (D) that covers the inland area for the reference climate is expected to shift into the temperate climate (C). The coefficient of variation (CVs) in the temperature ensemble is particularly low for the southern coast of the Korean peninsula, and accordingly, a high possibility of the shifting climatic zone of the coast is predicted. This research was supported by a grant (MOIS-DP-2015-05) of Disaster Prediction and Mitigation Technology Development Program funded by Ministry of Interior and Safety (MOIS, Korea).Keywords: MME, North Korea, Koppen–Geiger, climatic zones, coefficient of variation, CV
Procedia PDF Downloads 111564 Real Time Classification of Political Tendency of Twitter Spanish Users based on Sentiment Analysis
Authors: Marc Solé, Francesc Giné, Magda Valls, Nina Bijedic
Abstract:
What people say on social media has turned into a rich source of information to understand social behavior. Specifically, the growing use of Twitter social media for political communication has arisen high opportunities to know the opinion of large numbers of politically active individuals in real time and predict the global political tendencies of a specific country. It has led to an increasing body of research on this topic. The majority of these studies have been focused on polarized political contexts characterized by only two alternatives. Unlike them, this paper tackles the challenge of forecasting Spanish political trends, characterized by multiple political parties, by means of analyzing the Twitters Users political tendency. According to this, a new strategy, named Tweets Analysis Strategy (TAS), is proposed. This is based on analyzing the users tweets by means of discovering its sentiment (positive, negative or neutral) and classifying them according to the political party they support. From this individual political tendency, the global political prediction for each political party is calculated. In order to do this, two different strategies for analyzing the sentiment analysis are proposed: one is based on Positive and Negative words Matching (PNM) and the second one is based on a Neural Networks Strategy (NNS). The complete TAS strategy has been performed in a Big-Data environment. The experimental results presented in this paper reveal that NNS strategy performs much better than PNM strategy to analyze the tweet sentiment. In addition, this research analyzes the viability of the TAS strategy to obtain the global trend in a political context make up by multiple parties with an error lower than 23%.Keywords: political tendency, prediction, sentiment analysis, Twitter
Procedia PDF Downloads 238563 Micro-Transformation Strategy Of Residential Transportation Space Based On The Demand Of Residents: Taking A Residential District In Wuhan, China As An Example
Abstract:
With the acceleration of urbanization and motorization in China, the scale of cities and the travel distance of residents are constantly expanding, and the number of cars is continuously increasing, so the urban traffic problem is more and more serious. Traffic congestion, environmental pollution, energy consumption, travel safety and direct interference between traffic and other urban activities are increasingly prominent problems brought about by motorized development. This not only has a serious impact on the lives of the residents but also has a major impact on the healthy development of the city. The paper found that, in order to solve the development of motorization, a number of problems will arise; urban planning and traffic planning and design in residential planning often take into account the development of motorized traffic but neglects the demand for street life. This kind of planning has resulted in the destruction of the traditional communication space of the residential area, the pollution of noise and exhaust gas, and the potential safety risks of the residential area, which has disturbed the previously quiet and comfortable life of the residential area, resulting in the inconvenience of residents' life and the loss of street vitality. Based on these facts, this paper takes a residential area in Wuhan as the research object, through the actual investigation and research, from the perspective of micro-transformation analysis, combined with the concept of traffic micro-reconstruction governance. And research puts forward the residential traffic optimization strategies such as strengthening the interaction and connection between the residential area and the urban street system, street traffic classification and organization.Keywords: micro-transformation, residential traffic, residents demand, traffic microcirculation
Procedia PDF Downloads 116562 Milk Protein Genetic Variation and Haplotype Structure in Sudanse Indigenous Dairy Zebu Cattle
Authors: Ammar Said Ahmed, M. Reissmann, R. Bortfeldt, G. A. Brockmann
Abstract:
Milk protein genetic variants are of interest for characterizing domesticated mammalian species and breeds, and for studying associations with economic traits. The aim of this work was to analyze milk protein genetic variation in the Sudanese native cattle breeds, which have been gradually declining in numbers over the last years due to the breed substitution, and indiscriminate crossbreeding. The genetic variation at three milk protein genes αS1-casein (CSN1S1), αS2-casein (CSN1S2) and ƙ-casein (CSN3) was investigated in 250 animals belonging to five Bos indicus cattle breeds of Sudan (Butana, Kenana, White-nile, Erashy and Elgash). Allele specific primers were designed for five SNPs determine the CSN1S1 variants B and C, the CSN1S2 variants A and B, the CSN3 variants A, B and H. Allele, haplotype frequencies and genetic distances (D) were calculated and the phylogenetic tree was constructed. All breeds were found to be polymorphic for the studied genes. The CSN1S1*C variant was found very frequently (>0.63) in all analyzed breeds with highest frequency (0.82) in White-nile cattle. The CSN1S2*A variant (0.77) and CSN3*A variant (0.79) had highest frequency in Kenana cattle. Eleven haplotypes in casein gene cluster were inferred. Six of all haplotypes occurred in all breeds with remarkably deferent frequencies. The estimated D ranged from 0.004 to 0.049. The most distant breeds were White-nile and Kenana (D 0.0479). The results presented contribute to the genetic knowledge of indigenous cattle and can be used for proper definition and classification of the Sudanese cattle breeds as well as breeding, utilization, and potential development of conservation strategies for local breeds.Keywords: milk protein, genetic variation, casein haplotype, Bos indicus
Procedia PDF Downloads 436561 Control for Fluid Flow Behaviours of Viscous Fluids and Heat Transfer in Mini-Channel: A Case Study Using Numerical Simulation Method
Authors: Emmanuel Ophel Gilbert, Williams Speret
Abstract:
The control for fluid flow behaviours of viscous fluids and heat transfer occurrences within heated mini-channel is considered. Heat transfer and flow characteristics of different viscous liquids, such as engine oil, automatic transmission fluid, one-half ethylene glycol, and deionized water were numerically analyzed. Some mathematical applications such as Fourier series and Laplace Z-Transforms were employed to ascertain the behaviour-wave like structure of these each viscous fluids. The steady, laminar flow and heat transfer equations are reckoned by the aid of numerical simulation technique. Further, this numerical simulation technique is endorsed by using the accessible practical values in comparison with the anticipated local thermal resistances. However, the roughness of this mini-channel that is one of the physical limitations was also predicted in this study. This affects the frictional factor. When an additive such as tetracycline was introduced in the fluid, the heat input was lowered, and this caused pro rata effect on the minor and major frictional losses, mostly at a very minute Reynolds number circa 60-80. At this ascertained lower value of Reynolds numbers, there exists decrease in the viscosity and minute frictional losses as a result of the temperature of these viscous liquids been increased. It is inferred that the three equations and models are identified which supported the numerical simulation via interpolation and integration of the variables extended to the walls of the mini-channel, yields the utmost reliance for engineering and technology calculations for turbulence impacting jets in the near imminent age. Out of reasoning with a true equation that could support this control for the fluid flow, Navier-stokes equations were found to tangential to this finding. Though, other physical factors with respect to these Navier-stokes equations are required to be checkmated to avoid uncertain turbulence of the fluid flow. This paradox is resolved within the framework of continuum mechanics using the classical slip condition and an iteration scheme via numerical simulation method that takes into account certain terms in the full Navier-Stokes equations. However, this resulted in dropping out in the approximation of certain assumptions. Concrete questions raised in the main body of the work are sightseen further in the appendices.Keywords: frictional losses, heat transfer, laminar flow, mini-channel, number simulation, Reynolds number, turbulence, viscous fluids
Procedia PDF Downloads 176560 Crop Leaf Area Index (LAI) Inversion and Scale Effect Analysis from Unmanned Aerial Vehicle (UAV)-Based Hyperspectral Data
Authors: Xiaohua Zhu, Lingling Ma, Yongguang Zhao
Abstract:
Leaf Area Index (LAI) is a key structural characteristic of crops and plays a significant role in precision agricultural management and farmland ecosystem modeling. However, LAI retrieved from different resolution data contain a scaling bias due to the spatial heterogeneity and model non-linearity, that is, there is scale effect during multi-scale LAI estimate. In this article, a typical farmland in semi-arid regions of Chinese Inner Mongolia is taken as the study area, based on the combination of PROSPECT model and SAIL model, a multiple dimensional Look-Up-Table (LUT) is generated for multiple crops LAI estimation from unmanned aerial vehicle (UAV) hyperspectral data. Based on Taylor expansion method and computational geometry model, a scale transfer model considering both difference between inter- and intra-class is constructed for scale effect analysis of LAI inversion over inhomogeneous surface. The results indicate that, (1) the LUT method based on classification and parameter sensitive analysis is useful for LAI retrieval of corn, potato, sunflower and melon on the typical farmland, with correlation coefficient R2 of 0.82 and root mean square error RMSE of 0.43m2/m-2. (2) The scale effect of LAI is becoming obvious with the decrease of image resolution, and maximum scale bias is more than 45%. (3) The scale effect of inter-classes is higher than that of intra-class, which can be corrected efficiently by the scale transfer model established based Taylor expansion and Computational geometry. After corrected, the maximum scale bias can be reduced to 1.2%.Keywords: leaf area index (LAI), scale effect, UAV-based hyperspectral data, look-up-table (LUT), remote sensing
Procedia PDF Downloads 440559 Performance Evaluation of Routing Protocol in Cognitive Radio with Multi Technological Environment
Authors: M. Yosra, A. Mohamed, T. Sami
Abstract:
Over the past few years, mobile communication technologies have seen significant evolution. This fact promoted the implementation of many systems in a multi-technological setting. From one system to another, the Quality of Service (QoS) provided to mobile consumers gets better. The growing number of normalized standards extends the available services for each consumer, moreover, most of the available radio frequencies have already been allocated, such as 3G, Wifi, Wimax, and LTE. A study by the Federal Communications Commission (FCC) found that certain frequency bands are partially occupied in particular locations and times. So, the idea of Cognitive Radio (CR) is to share the spectrum between a primary user (PU) and a secondary user (SU). The main objective of this spectrum management is to achieve a maximum rate of exploitation of the radio spectrum. In general, the CR can greatly improve the quality of service (QoS) and improve the reliability of the link. The problem will reside in the possibility of proposing a technique to improve the reliability of the wireless link by using the CR with some routing protocols. However, users declared that the links were unreliable and that it was an incompatibility with QoS. In our case, we choose the QoS parameter "bandwidth" to perform a supervised classification. In this paper, we propose a comparative study between some routing protocols, taking into account the variation of different technologies on the existing spectral bandwidth like 3G, WIFI, WIMAX, and LTE. Due to the simulation results, we observe that LTE has significantly higher availability bandwidth compared with other technologies. The performance of the OLSR protocol is better than other on-demand routing protocols (DSR, AODV and DSDV), in LTE technology because of the proper receiving of packets, less packet drop and the throughput. Numerous simulations of routing protocols have been made using simulators such as NS3.Keywords: cognitive radio, multi technology, network simulator (NS3), routing protocol
Procedia PDF Downloads 63558 Author Profiling: Prediction of Learners’ Gender on a MOOC Platform Based on Learners’ Comments
Authors: Tahani Aljohani, Jialin Yu, Alexandra. I. Cristea
Abstract:
The more an educational system knows about a learner, the more personalised interaction it can provide, which leads to better learning. However, asking a learner directly is potentially disruptive, and often ignored by learners. Especially in the booming realm of MOOC Massive Online Learning platforms, only a very low percentage of users disclose demographic information about themselves. Thus, in this paper, we aim to predict learners’ demographic characteristics, by proposing an approach using linguistically motivated Deep Learning Architectures for Learner Profiling, particularly targeting gender prediction on a FutureLearn MOOC platform. Additionally, we tackle here the difficult problem of predicting the gender of learners based on their comments only – which are often available across MOOCs. The most common current approaches to text classification use the Long Short-Term Memory (LSTM) model, considering sentences as sequences. However, human language also has structures. In this research, rather than considering sentences as plain sequences, we hypothesise that higher semantic - and syntactic level sentence processing based on linguistics will render a richer representation. We thus evaluate, the traditional LSTM versus other bleeding edge models, which take into account syntactic structure, such as tree-structured LSTM, Stack-augmented Parser-Interpreter Neural Network (SPINN) and the Structure-Aware Tag Augmented model (SATA). Additionally, we explore using different word-level encoding functions. We have implemented these methods on Our MOOC dataset, which is the most performant one comparing with a public dataset on sentiment analysis that is further used as a cross-examining for the models' results.Keywords: deep learning, data mining, gender predication, MOOCs
Procedia PDF Downloads 147557 Further Evidence for the Existence of Broiler Chicken PFN (Pale, Firm and Non-Exudative Meat) and PSE (Pale, Soft and Exudative) in Brazilian Commercial Flocks
Authors: Leila M. Carvalho, Maria Erica S. Oliveira, Arnoud C. Neto, Elza I. Ida, Massami Shimokomaki, Marta S. Madruga
Abstract:
The quality of broiler breast meat is changing as a result of the continuing emphasis on genetic selection for a more efficient meat production. Breast meat has been classified as PSE (pale, soft, exudative), DFD (dark, firm, dry) and normal color meat, and recently a third group has emerged: the so-called PFN (pale, firm, non-exudative) meat. This classification was based on pH, color and functional properties. The aim of this work was to confirm the existence of PFN and PSE meat by biochemical characterization and functional properties. Twenty four hours of refrigerated fillet, Pectoralis major, m. samples (n= 838) were taken from Cobb flocks 42-48 days old, obtained in Northeastern Brazil tropical region, the Northeastern, considered to have only dry and wet seasons. Color (L*), pH, water holding capacity (WHC), values were evaluated and compared with PSE group samples. These samples were classified as Normal (46Keywords: broiler breast meat, funcional properties, PFN, PSE
Procedia PDF Downloads 249556 Differential Response of Cellular Antioxidants and Proteome Expression to Salt, Cadmium and Their Combination in Spinach (Spinacia oleracea)
Authors: Rita Bagheri, Javed Ahmed, Humayra Bashir, M. Irfan Qureshi
Abstract:
Agriculture lands suffer from a combination of stresses such as salinity and metal contamination including cadmium at the same time. Under such condition of multiple stresses, plant may exhibit unique responses different from the stress occurring individually. Thus, it would be interesting to investigate that how plant respond to combined stress at level of antioxidants and proteome expression, and identifying the proteins which are involved in imparting stress tolerance. With an approach of comparative proteomics and antioxidant analysis, present study investigates the response of Spinacia oleracea to salt (NaCl), cadmium (Cd), and their combination (NaCl+Cd) stress. Two-dimensional gel electrophoresis was used for resolving leaf proteome, and proteins of interest were identified using PDQuest software. A number of proteins expressed differentially, those indicated towards their roles in imparting stress tolerance, were digested by trypsin and analyzed on mass spectrometer for peptide mass fingerprinting (PMF). Data signals were then matched with protein databases using MASCOT. Results show that NaCl, Cd and both together (NaCl+Cd) induce oxidative stress which was highest in combined stress of Cd+NaCl. Correspondingly, the activities of enzymatic antioxidants viz., SOD, APX, GR and CAT, and non-enzymatic antioxidants had highest changes under combined stress compares to single stress over their respective controls. Among the identified proteins, several interesting proteins were identified that may be have role in Spinacia oleracia tolerance in individual and combinatorial stress of salt and cadmium. The functional classification of identified proteins indicates the importance and necessity of keeping higher ratio of defence and disease responsive proteins.Keywords: Spinacia oleracea, Cd, salinity, proteomics, antioxidants, combinatorial stress
Procedia PDF Downloads 382555 Optimizing Production Yield Through Process Parameter Tuning Using Deep Learning Models: A Case Study in Precision Manufacturing
Authors: Tolulope Aremu
Abstract:
This paper is based on the idea of using deep learning methodology for optimizing production yield by tuning a few key process parameters in a manufacturing environment. The study was explicitly on how to maximize production yield and minimize operational costs by utilizing advanced neural network models, specifically Long Short-Term Memory and Convolutional Neural Networks. These models were implemented using Python-based frameworks—TensorFlow and Keras. The targets of the research are the precision molding processes in which temperature ranges between 150°C and 220°C, the pressure ranges between 5 and 15 bar, and the material flow rate ranges between 10 and 50 kg/h, which are critical parameters that have a great effect on yield. A dataset of 1 million production cycles has been considered for five continuous years, where detailed logs are present showing the exact setting of parameters and yield output. The LSTM model would model time-dependent trends in production data, while CNN analyzed the spatial correlations between parameters. Models are designed in a supervised learning manner. For the model's loss, an MSE loss function is used, optimized through the Adam optimizer. After running a total of 100 training epochs, 95% accuracy was achieved by the models recommending optimal parameter configurations. Results indicated that with the use of RSM and DOE traditional methods, there was an increase in production yield of 12%. Besides, the error margin was reduced by 8%, hence consistent quality products from the deep learning models. The monetary value was annually around $2.5 million, the cost saved from material waste, energy consumption, and equipment wear resulting from the implementation of optimized process parameters. This system was deployed in an industrial production environment with the help of a hybrid cloud system: Microsoft Azure, for data storage, and the training and deployment of their models were performed on Google Cloud AI. The functionality of real-time monitoring of the process and automatic tuning of parameters depends on cloud infrastructure. To put it into perspective, deep learning models, especially those employing LSTM and CNN, optimize the production yield by fine-tuning process parameters. Future research will consider reinforcement learning with a view to achieving further enhancement of system autonomy and scalability across various manufacturing sectors.Keywords: production yield optimization, deep learning, tuning of process parameters, LSTM, CNN, precision manufacturing, TensorFlow, Keras, cloud infrastructure, cost saving
Procedia PDF Downloads 29554 Study of Parking Demand for Offices – Case Study: Kolkata
Authors: Sanghamitra Roy
Abstract:
In recent times, India has experienced the phenomenal rise in the number of registered vehicles and vehicular trips, particularly intra-city trips in most of its urban areas. The increase in vehicle ownership and use have increased parking demand immensely and accommodating the same is now a matter of big concern. Most cities do not have adequate off-street parking facilities thus forcing people to park on the streets. This has resulted in decreased carrying capacity, decreased traffic speed, increased congestion, and increased environmental problems. While integrated multi-modal transportation system is the answer to such problems, parking issues will continue to exist. In Kolkata, only 6.4% land is devoted for roads. The consequences of this huge crunch in road spaces coupled with increased parking demand are severe particularly in the CBD and major commercial areas, making the role of off-street parking facilities in Kolkata even more critical. To meaningfully address parking issues, it is important to identify the factors that influence parking demand so that it can be assessed and comprehensive parking policies and plans for the city can be formulated. This paper aims at identifying the factors that contribute towards parking demand for offices in Kolkata and their degree of correlation with parking demand. The study is limited to home-to-work trips located within Kolkata Municipal Corporation (KMC) where parking related issues are most pronounced. The data for the study is collected through personal interviews, questionnaires and direct observations from offices across the wards of KMC. SPSS is used for classification of the data and analyses of the same. The findings of this study will help in re-assessment of the parking requirements specified in The Kolkata Municipal Corporation Building Rules as a step towards alleviating parking related issues in the city.Keywords: building rules, office spaces, parking demand, urbanization
Procedia PDF Downloads 317553 Social Media Data Analysis for Personality Modelling and Learning Styles Prediction Using Educational Data Mining
Authors: Srushti Patil, Preethi Baligar, Gopalkrishna Joshi, Gururaj N. Bhadri
Abstract:
In designing learning environments, the instructional strategies can be tailored to suit the learning style of an individual to ensure effective learning. In this study, the information shared on social media like Facebook is being used to predict learning style of a learner. Previous research studies have shown that Facebook data can be used to predict user personality. Users with a particular personality exhibit an inherent pattern in their digital footprint on Facebook. The proposed work aims to correlate the user's’ personality, predicted from Facebook data to the learning styles, predicted through questionnaires. For Millennial learners, Facebook has become a primary means for information sharing and interaction with peers. Thus, it can serve as a rich bed for research and direct the design of learning environments. The authors have conducted this study in an undergraduate freshman engineering course. Data from 320 freshmen Facebook users was collected. The same users also participated in the learning style and personality prediction survey. The Kolb’s Learning style questionnaires and Big 5 personality Inventory were adopted for the survey. The users have agreed to participate in this research and have signed individual consent forms. A specific page was created on Facebook to collect user data like personal details, status updates, comments, demographic characteristics and egocentric network parameters. This data was captured by an application created using Python program. The data captured from Facebook was subjected to text analysis process using the Linguistic Inquiry and Word Count dictionary. An analysis of the data collected from the questionnaires performed reveals individual student personality and learning style. The results obtained from analysis of Facebook, learning style and personality data were then fed into an automatic classifier that was trained by using the data mining techniques like Rule-based classifiers and Decision trees. This helps to predict the user personality and learning styles by analysing the common patterns. Rule-based classifiers applied for text analysis helps to categorize Facebook data into positive, negative and neutral. There were totally two models trained, one to predict the personality from Facebook data; another one to predict the learning styles from the personalities. The results show that the classifier model has high accuracy which makes the proposed method to be a reliable one for predicting the user personality and learning styles.Keywords: educational data mining, Facebook, learning styles, personality traits
Procedia PDF Downloads 231552 Computational Intelligence and Machine Learning for Urban Drainage Infrastructure Asset Management
Authors: Thewodros K. Geberemariam
Abstract:
The rapid physical expansion of urbanization coupled with aging infrastructure presents a unique decision and management challenges for many big city municipalities. Cities must therefore upgrade and maintain the existing aging urban drainage infrastructure systems to keep up with the demands. Given the overall contribution of assets to municipal revenue and the importance of infrastructure to the success of a livable city, many municipalities are currently looking for a robust and smart urban drainage infrastructure asset management solution that combines management, financial, engineering and technical practices. This robust decision-making shall rely on sound, complete, current and relevant data that enables asset valuation, impairment testing, lifecycle modeling, and forecasting across the multiple asset portfolios. On this paper, predictive computational intelligence (CI) and multi-class machine learning (ML) coupled with online, offline, and historical record data that are collected from an array of multi-parameter sensors are used for the extraction of different operational and non-conforming patterns hidden in structured and unstructured data to determine and produce actionable insight on the current and future states of the network. This paper aims to improve the strategic decision-making process by identifying all possible alternatives; evaluate the risk of each alternative, and choose the alternative most likely to attain the required goal in a cost-effective manner using historical and near real-time urban drainage infrastructure data for urban drainage infrastructures assets that have previously not benefited from computational intelligence and machine learning advancements.Keywords: computational intelligence, machine learning, urban drainage infrastructure, machine learning, classification, prediction, asset management space
Procedia PDF Downloads 152551 C-eXpress: A Web-Based Analysis Platform for Comparative Functional Genomics and Proteomics in Human Cancer Cell Line, NCI-60 as an Example
Authors: Chi-Ching Lee, Po-Jung Huang, Kuo-Yang Huang, Petrus Tang
Abstract:
Background: Recent advances in high-throughput research technologies such as new-generation sequencing and multi-dimensional liquid chromatography makes it possible to dissect the complete transcriptome and proteome in a single run for the first time. However, it is almost impossible for many laboratories to handle and analysis these “BIG” data without the support from a bioinformatics team. We aimed to provide a web-based analysis platform for users with only limited knowledge on bio-computing to study the functional genomics and proteomics. Method: We use NCI-60 as an example dataset to demonstrate the power of the web-based analysis platform and data delivering system: C-eXpress takes a simple text file that contain the standard NCBI gene or protein ID and expression levels (rpkm or fold) as input file to generate a distribution map of gene/protein expression levels in a heatmap diagram organized by color gradients. The diagram is hyper-linked to a dynamic html table that allows the users to filter the datasets based on various gene features. A dynamic summary chart is generated automatically after each filtering process. Results: We implemented an integrated database that contain pre-defined annotations such as gene/protein properties (ID, name, length, MW, pI); pathways based on KEGG and GO biological process; subcellular localization based on GO cellular component; functional classification based on GO molecular function, kinase, peptidase and transporter. Multiple ways of sorting of column and rows is also provided for comparative analysis and visualization of multiple samples.Keywords: cancer, visualization, database, functional annotation
Procedia PDF Downloads 618550 Managing Pseudoangiomatous Stromal Hyperplasia Appropriately and Safely: A Retrospective Case Series Review
Authors: C. M. Williams, R. English, P. King, I. M. Brown
Abstract:
Introduction: Pseudoangiomatous Stromal Hyperplasia (PASH) is a benign fibrous proliferation of breast stroma affecting predominantly premenopausal women with no significant increased risk of breast cancer. Informal recommendations for management have continued to evolve over recent years from surgical excision to observation, although there are no specific national guidelines. This study assesses the safety of a non-surgical approach to PASH management by review of cases at a single centre. Methods: Retrospective case series review (January 2011 – August 2016) was conducted on consecutive PASH cases. Diagnostic classification (clinical, radiological and histological), management outcomes, and breast cancer incidence were recorded. Results: 43 patients were followed up for median of 25 months (3-64) with 75% symptomatic at presentation. 12% of cases (n=5) had a radiological score (BIRADS MMG or US) ≥ 4 of which 3 were confirmed malignant. One further malignancy was detected and proven radiologically occult and contralateral. No patients were diagnosed with a malignancy during follow-up. Treatment evolved from 67% surgical in 2011 to 33% in 2016. Conclusions: The management of PASH has transitioned in line with other published experience. The preliminary findings suggest this appears safe with no evidence of missed malignancies; however, longer follow up is required to confirm long-term safety. Recommendations: PASH with suspicious radiological findings ( ≥ U4/R4) warrants multidisciplinary discussion for excision. In the absence of histological or radiological suspicion of malignancy, PASH can be safely managed without surgery.Keywords: benign breast disease, conservative management, malignancy, pseudoangiomatous stromal hyperplasia, surgical excision
Procedia PDF Downloads 132549 A Comparative Study of Dengue Fever in Taiwan and Singapore Based on Open Data
Authors: Wei Wen Yang, Emily Chia Yu Su
Abstract:
Dengue fever is a mosquito-borne tropical infectious disease caused by the dengue virus. After infection, symptoms usually start from three to fourteen days. Dengue virus may cause a high fever and at least two of the following symptoms, severe headache, severe eye pain, joint pains, muscle or bone pain, vomiting, feature skin rash, and mild bleeding manifestation. In addition, recovery will take at least two to seven days. Dengue fever has rapidly spread in tropical and subtropical areas in recent years. Several phenomena around the world such as global warming, urbanization, and international travel are the main reasons in boosting the spread of dengue. In Taiwan, epidemics occur annually, especially during summer and fall seasons. On the other side, Singapore government also has announced the amounts number of dengue cases spreading in Singapore. As the serious epidemic of dengue fever outbreaks in Taiwan and Singapore, countries around the Asia-Pacific region are becoming high risks of susceptible to the outbreaks and local hub of spreading the virus. To improve public safety and public health issues, firstly, we are going to use Microsoft Excel and SAS EG to do data preprocessing. Secondly, using support vector machines and decision trees builds predict model, and analyzes the infectious cases between Taiwan and Singapore. By comparing different factors causing vector mosquito from model classification and regression, we can find similar spreading patterns where the disease occurred most frequently. The result can provide sufficient information to predict the future dengue infection outbreaks and control the diffusion of dengue fever among countries.Keywords: dengue fever, Taiwan, Singapore, Aedes aegypti
Procedia PDF Downloads 234548 An Extensive Review of Drought Indices
Authors: Shamsulhaq Amin
Abstract:
Drought can arise from several hydrometeorological phenomena that result in insufficient precipitation, soil moisture, and surface and groundwater flow, leading to conditions that are considerably drier than the usual water content or availability. Drought is often assessed using indices that are associated with meteorological, agricultural, and hydrological phenomena. In order to effectively handle drought disasters, it is essential to accurately determine the kind, intensity, and extent of the drought using drought characterization. This information is critical for managing the drought before, during, and after the rehabilitation process. Over a hundred drought assessments have been created in literature to evaluate drought disasters, encompassing a range of factors and variables. Some models utilise solely hydrometeorological drivers, while others employ remote sensing technology, and some incorporate a combination of both. Comprehending the entire notion of drought and taking into account drought indices along with their calculation processes are crucial for researchers in this discipline. Examining several drought metrics in different studies requires additional time and concentration. Hence, it is crucial to conduct a thorough examination of approaches used in drought indices in order to identify the most straightforward approach to avoid any discrepancies in numerous scientific studies. In case of practical application in real-world, categorizing indices relative to their usage in meteorological, agricultural, and hydrological phenomena might help researchers maximize their efficiency. Users have the ability to explore different indexes at the same time, allowing them to compare the convenience of use and evaluate the benefits and drawbacks of each. Moreover, certain indices exhibit interdependence, which enhances comprehension of their connections and assists in making informed decisions about their suitability in various scenarios. This study provides a comprehensive assessment of various drought indices, analysing their types and computation methodologies in a detailed and systematic manner.Keywords: drought classification, drought severity, drought indices, agriculture, hydrological
Procedia PDF Downloads 41547 Early Gastric Cancer Prediction from Diet and Epidemiological Data Using Machine Learning in Mizoram Population
Authors: Brindha Senthil Kumar, Payel Chakraborty, Senthil Kumar Nachimuthu, Arindam Maitra, Prem Nath
Abstract:
Gastric cancer is predominantly caused by demographic and diet factors as compared to other cancer types. The aim of the study is to predict Early Gastric Cancer (ECG) from diet and lifestyle factors using supervised machine learning algorithms. For this study, 160 healthy individual and 80 cases were selected who had been followed for 3 years (2016-2019), at Civil Hospital, Aizawl, Mizoram. A dataset containing 11 features that are core risk factors for the gastric cancer were extracted. Supervised machine algorithms: Logistic Regression, Naive Bayes, Support Vector Machine (SVM), Multilayer perceptron, and Random Forest were used to analyze the dataset using Python Jupyter Notebook Version 3. The obtained classified results had been evaluated using metrics parameters: minimum_false_positives, brier_score, accuracy, precision, recall, F1_score, and Receiver Operating Characteristics (ROC) curve. Data analysis results showed Naive Bayes - 88, 0.11; Random Forest - 83, 0.16; SVM - 77, 0.22; Logistic Regression - 75, 0.25 and Multilayer perceptron - 72, 0.27 with respect to accuracy and brier_score in percent. Naive Bayes algorithm out performs with very low false positive rates as well as brier_score and good accuracy. Naive Bayes algorithm classification results in predicting ECG showed very satisfactory results using only diet cum lifestyle factors which will be very helpful for the physicians to educate the patients and public, thereby mortality of gastric cancer can be reduced/avoided with this knowledge mining work.Keywords: Early Gastric cancer, Machine Learning, Diet, Lifestyle Characteristics
Procedia PDF Downloads 161546 The Relationship of Socioeconomic Status and Levels of Delinquency among Senior High School Students with Secured Attachment to Their Mothers
Authors: Aldrin Avergas, Quennie Mariel Peñaranda, Niña Karen San Miguel, Alexis Katrina Agustin, Peralta Xusha Mae, Maria Luisa Sison
Abstract:
The research is entitled “The Relationship of Socioeconomic Status and Levels of Delinquency among Senior High School Students with Secured Attachment to their Mothers”. The researchers had explored the relationship between socioeconomic status and delinquent tendencies among grade 11 students. The objective of the research is to discover if delinquent behavior will have a relationship with the current socio-economic status of an adolescent student having a warm relationship with their mothers. The researchers utilized three questionnaires that would measure the three variables of the study, namely: (1) 1SEC 2012: The New Philippines Socioeconomic Classification System was used to show the current socioeconomic status of the respondents, (2) Self-Reported Delinquency – Problem Behavior Frequency Scale was utilized to determine the individual's frequency in engaging to delinquent behavior, and (3) Inventory of Parent and Peer Attachment Revised (IPPA-R) was used to determine the attachment style of the respondents. The researchers utilized a quantitative research design, specifically correlation research. The study concluded that there is no significant relationship between socioeconomic status and academic delinquency despite the fact that these participants had secured attachment to their mother hence this research implies that delinquency is not just a problem for students belonging in the lower socio-economic status and that even having a warm and close relationship with their mothers is not sufficient enough for these students to completely be free from engaging in delinquent acts. There must be other factors (such as peer pressure, emotional quotient, self-esteem or etc.) that are might be contributing to delinquent behaviors.Keywords: adolescents, delinquency, high school students, secured attachment style, socioeconomic status
Procedia PDF Downloads 186545 Unlocking the Genetic Code: Exploring the Potential of DNA Barcoding for Biodiversity Assessment
Authors: Mohammed Ahmed Ahmed Odah
Abstract:
DNA barcoding is a crucial method for assessing and monitoring species diversity amidst escalating threats to global biodiversity. The author explores DNA barcoding's potential as a robust and reliable tool for biodiversity assessment. It begins with a comprehensive review of existing literature, delving into the theoretical foundations, methodologies and applications of DNA barcoding. The suitability of various DNA regions, like the COI gene, as universal barcodes is extensively investigated. Additionally, the advantages and limitations of different DNA sequencing technologies and bioinformatics tools are evaluated within the context of DNA barcoding. To evaluate the efficacy of DNA barcoding, diverse ecosystems, including terrestrial, freshwater and marine habitats, are sampled. Extracted DNA from collected specimens undergoes amplification and sequencing of the target barcode region. Comparison of the obtained DNA sequences with reference databases allows for the identification and classification of the sampled organisms. Findings demonstrate that DNA barcoding accurately identifies species, even in cases where morphological identification proves challenging. Moreover, it sheds light on cryptic and endangered species, aiding conservation efforts. The author also investigates patterns of genetic diversity and evolutionary relationships among different taxa through the analysis of genetic data. This research contributes to the growing knowledge of DNA barcoding and its applicability for biodiversity assessment. The advantages of this approach, such as speed, accuracy and cost-effectiveness, are highlighted, along with areas for improvement. By unlocking the genetic code, DNA barcoding enhances our understanding of biodiversity, supports conservation initiatives and informs evidence-based decision-making for the sustainable management of ecosystems.Keywords: DNA barcoding, biodiversity assessment, genetic code, species identification, taxonomic resolution, next-generation sequencing
Procedia PDF Downloads 24544 Management of Interdependence in Manufacturing Networks
Authors: Atour Taghipour
Abstract:
In the real world each manufacturing company is an independent business unit. These business units are linked to each other through upstream and downstream linkages. The management of these linkages is called coordination which, could be considered as a difficult engineering task. The degree of difficulty of coordination depends on the type and the nature of information exchanged between partners as well as the structure of relationship from mutual to the network structure. The literature of manufacturing systems comprises a wide range of varieties of methods and approaches of coordination. In fact, two main streams of research can be distinguished: central coordination versus decentralized coordination. In the centralized systems a high degree of information exchanges is required. The high degree of information exchanges sometimes leads to difficulties when independent members do not want to share information. In order to address these difficulties, decentralized approaches of coordination of operations planning decisions based on some minimal information sharing have been proposed in many academic disciplines. This paper first proposes a framework of analysis in order to analyze the proposed approaches in the literature, based on this framework which includes the similarities between approaches we categorize the existing approaches. This classification can be used as a research map for future researches. The result of our paper highlights several opportunities for future research. First, it is proposed to develop more dynamic and stochastic mechanisms of planning coordination of manufacturing units. Second, in order to exploit the complementarities of approaches proposed by diverse science discipline, we propose to integrate the techniques of coordination. Finally, based on our approach we proposed to develop coordination standards to guaranty both the complementarity of these approaches as well as the freedom of companies to adopt any planning tools.Keywords: network coordination, manufacturing, operations planning, supply chain
Procedia PDF Downloads 281543 Medical Diagnosis of Retinal Diseases Using Artificial Intelligence Deep Learning Models
Authors: Ethan James
Abstract:
Over one billion people worldwide suffer from some level of vision loss or blindness as a result of progressive retinal diseases. Many patients, particularly in developing areas, are incorrectly diagnosed or undiagnosed whatsoever due to unconventional diagnostic tools and screening methods. Artificial intelligence (AI) based on deep learning (DL) convolutional neural networks (CNN) have recently gained a high interest in ophthalmology for its computer-imaging diagnosis, disease prognosis, and risk assessment. Optical coherence tomography (OCT) is a popular imaging technique used to capture high-resolution cross-sections of retinas. In ophthalmology, DL has been applied to fundus photographs, optical coherence tomography, and visual fields, achieving robust classification performance in the detection of various retinal diseases including macular degeneration, diabetic retinopathy, and retinitis pigmentosa. However, there is no complete diagnostic model to analyze these retinal images that provide a diagnostic accuracy above 90%. Thus, the purpose of this project was to develop an AI model that utilizes machine learning techniques to automatically diagnose specific retinal diseases from OCT scans. The algorithm consists of neural network architecture that was trained from a dataset of over 20,000 real-world OCT images to train the robust model to utilize residual neural networks with cyclic pooling. This DL model can ultimately aid ophthalmologists in diagnosing patients with these retinal diseases more quickly and more accurately, therefore facilitating earlier treatment, which results in improved post-treatment outcomes.Keywords: artificial intelligence, deep learning, imaging, medical devices, ophthalmic devices, ophthalmology, retina
Procedia PDF Downloads 181