Search results for: source code
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5737

Search results for: source code

5527 A Qualitative Evidence of the Markedness of Code Switching during Commercial Bank Service Encounters in Ìbàdàn Metropolis

Authors: A. Robbin

Abstract:

In a multilingual setting like Nigeria, the success of service encounters is enhanced by the use of a language that ensures the linguistic and persuasive demands of the interlocutors. This study examined motivations for code switching as a negotiation strategy in bank-hall desk service encounters in Ìbàdàn metropolis using Myers-Scotton’s exploration on markedness in language use. The data consisted of transcribed audio recording of bank-hall service encounters, and direct observation of bank interactions in two purposively sampled commercial banks in Ìbàdàn metropolis. The data was subjected to descriptive linguistic analysis using Myers Scotton’s Markedness Model.  Findings reveal that code switching is frequently employed during different stages of service encounter: greeting, transaction and closing to fulfil relational, bargaining and referential functions. Bank staff and customers code switch to make unmarked, marked and explanatory choices. A strategy used to identify with customer’s cultural affiliation, close status gap, and appeal to begrudged customer; or as an explanatory choice with non-literate customers for ease of communication. Bankers select English to maintain customers’ perceptions of prestige which is retained or diverged from depending on their linguistic preference or ability.  Yoruba is seen as an efficient negotiation strategy with both bankers and their customers, making choices within conversation to achieve desired conversational and functional aims.

Keywords: banking, bilingualism, code-switching, markedness, service encounter

Procedia PDF Downloads 178
5526 Myanmar Character Recognition Using Eight Direction Chain Code Frequency Features

Authors: Kyi Pyar Zaw, Zin Mar Kyu

Abstract:

Character recognition is the process of converting a text image file into editable and searchable text file. Feature Extraction is the heart of any character recognition system. The character recognition rate may be low or high depending on the extracted features. In the proposed paper, 25 features for one character are used in character recognition. Basically, there are three steps of character recognition such as character segmentation, feature extraction and classification. In segmentation step, horizontal cropping method is used for line segmentation and vertical cropping method is used for character segmentation. In the Feature extraction step, features are extracted in two ways. The first way is that the 8 features are extracted from the entire input character using eight direction chain code frequency extraction. The second way is that the input character is divided into 16 blocks. For each block, although 8 feature values are obtained through eight-direction chain code frequency extraction method, we define the sum of these 8 feature values as a feature for one block. Therefore, 16 features are extracted from that 16 blocks in the second way. We use the number of holes feature to cluster the similar characters. We can recognize the almost Myanmar common characters with various font sizes by using these features. All these 25 features are used in both training part and testing part. In the classification step, the characters are classified by matching the all features of input character with already trained features of characters.

Keywords: chain code frequency, character recognition, feature extraction, features matching, segmentation

Procedia PDF Downloads 290
5525 The Connection between the Schwartz Theory of Basic Values and Ethical Principles in Clinical Psychology

Authors: Matej Stritesky

Abstract:

The research deals with the connection between the Schwartz Theory of Basic Values and the ethical principles in psychology, on which the meta-code of ethics the European Federation of Psychological Associations is based. The research focuses on ethically problematic situations in clinical psychology in the Czech Republic. Based on the analysis of papers that identified ethically problematic situations faced by clinical psychologists, a questionnaire of ethically problematic situations in clinical psychology (EPSCP) was created for the purposes of the research. The questionnaire was created to represent situations that correspond to the 4 principles on which the meta-code of ethics the European Federation of Psychological Associations is based. The questionnaire EPSCP consists of descriptions of 32 situations that respondents evaluate on a scale from 1 (psychologist's behaviour is ethically perfectly fine) to 10 (psychologist's behaviour is ethically completely unacceptable). The EPSCP questionnaire, together with Schwartz's PVQ questionnaire, will be presented to 60 psychology students. The relationship between principles in clinical psychology and the values on Schwartz´s value continuum will be described using multidimensional scaling. A positive correlation is assumed between the higher-order value of openness to change and problematic ethical situations related to the principle of integrity; a positive correlation between the value of the higher order of self-transcendence and the principle of respect and responsibility; a positive correlation between the value of the higher order of conservation and the principle of competence; and negative correlation between the value of the higher order of ego strengthening and sensitivity to ethically problematic situations. The research also includes an experimental part. The first half of the students are presented with the code of ethics of the Czech Association of Clinical Psychologists before completing the questionnaires, and to the second half of the students is the code of ethics presented after completing the questionnaires. In addition to reading the code of ethics, students describe the three rules of the code of ethics that they consider most important and state why they chose these rules. The output of the experimental part will be to determine whether the presentation of the code of ethics leads to greater sensitivity to ethically problematic situations.

Keywords: clinical psychology, ethically problematic situations in clinical psychology, ethical principles in psychology, Schwartz theory of basic values

Procedia PDF Downloads 89
5524 Numerical Study of a Nanofluid in a Truncated Cone

Authors: B. Mahfoud, A. Bendjaghlouli

Abstract:

Natural convection is simulated in a truncated cone filled with nanofluid. Inclined and top walls have constant temperature where the heat source is located on the bottom wall of the conical container which is thermally insulated. A finite volume approach is used to solve the governing equations using the SIMPLE algorithm for different parameters such as Rayleigh number, inclination angle of inclined walls of the enclosure and heat source length. The results showed an enhancement in cooling system by using a nanofluid, when conduction regime is assisted. The inclination angle of inclined sidewall and heat source length affect the heat transfer rate and the maximum temperature.

Keywords: heat source, truncated cone, nanofluid, natural convection

Procedia PDF Downloads 286
5523 Cold Flow Investigation of Silicon Carbide Cylindrical Filter Element

Authors: Mohammad Alhajeri

Abstract:

This paper reports a computational fluid dynamics (CFD) investigation of cylindrical filter. Silicon carbide cylindrical filter elements have proven to be an effective mean of removing particulates to levels exceeding the new source performance standard. The CFD code is used here to understand the deposition process and the factors that affect the particles distribution over the filter element surface. Different approach cross flow velocity to filter face velocity ratios and different face velocities (ranging from 2 to 5 cm/s) are used in this study. Particles in the diameter range 1 to 100 microns are tracked through the domain. The radius of convergence (or the critical trajectory) is compared and plotted as a function of many parameters.

Keywords: filtration, CFD, CCF, hot gas filtration

Procedia PDF Downloads 438
5522 Simulation of Reflectometry in Alborz Tokamak

Authors: S. Kohestani, R. Amrollahi, P. Daryabor

Abstract:

Microwave diagnostics such as reflectometry are receiving growing attention in magnetic confinement fusionresearch. In order to obtain the better understanding of plasma confinement physics, more detailed measurements on density profile and its fluctuations might be required. A 2D full-wave simulation of ordinary mode propagation has been written in an effort to model effects seen in reflectometry experiment. The code uses the finite-difference-time-domain method with a perfectly-matched-layer absorption boundary to solve Maxwell’s equations.The code has been used to simulate the reflectometer measurement in Alborz Tokamak.

Keywords: reflectometry, simulation, ordinary mode, tokamak

Procedia PDF Downloads 395
5521 Heat Source Temperature for Centered Heat Source on Isotropic Plate with Lower Surface Forced Cooling Using Neural Network and Three Different Materials

Authors: Fadwa Haraka, Ahmad Elouatouati, Mourad Taha Janan

Abstract:

In this study, we propose a neural network based method in order to calculate the heat source temperature of isotropic plate with lower surface forced cooling. To validate the proposed model, the heat source temperatures values will be compared to the analytical method -variables separation- and finite element model. The mathematical simulation is done through 3D numerical simulation by COMSOL software considering three different materials: Aluminum, Copper, and Graphite. The proposed method will lead to a formulation of the heat source temperature based on the thermal and geometric properties of the base plate.

Keywords: thermal model, thermal resistance, finite element simulation, neural network

Procedia PDF Downloads 331
5520 Difference between 'HDR Ir-192 and Co-60 Sources' for High Dose Rate Brachytherapy Machine

Authors: Md Serajul Islam

Abstract:

High Dose Rate (HDR) Brachytherapy is used for cancer patients. In our country’s prospect, we are using only cervices and breast cancer treatment by using HDR. The air kerma rate in air at a reference distance of less than a meter from the source is the recommended quantity for the specification of gamma ray source Ir-192 in brachytherapy. The absorbed dose for the patients is directly proportional to the air kerma rate. Therefore the air kerma rate should be determined before the first use of the source on patients by qualified medical physicist who is independent from the source manufacturer. The air kerma rate will then be applied in the calculation of the dose delivered to patients in their planning systems. In practice, high dose rate (HDR) Ir-192 afterloader machines are mostly used in brachytherapy treatment. Currently, HDR-Co-60 increasingly comes into operation too. The essential advantage of the use of Co-60 sources is its longer half-life compared to Ir-192. The use of HDRCo-60 afterloading machines is also quite interesting for developing countries. This work describes the dosimetry at HDR afterloading machines according to the protocols IAEA-TECDOC-1274 (2002) with the nuclides Ir-192 and Co-60. We have used 3 different measurement methods (with a ring chamber, with a solid phantom and in free air and with a well chamber) in dependence of each of the protocols. We have shown that the standard deviations of the measured air kerma rate for the Co-60 source are generally larger than those of the Ir-192 source. The measurements with the well chamber had the lowest deviation from the certificate value. In all protocols and methods, the deviations stood for both nuclides by a maximum of about 1% for Ir-192 and 2.5% for Co-60-Sources respectively.

Keywords: Ir-192 source, cancer, patients, cheap treatment cost

Procedia PDF Downloads 201
5519 Clinch Process Simulation Using Diffuse Elements

Authors: Benzegaou Ali, Brani Benabderrahmane

Abstract:

This work describes a numerical study of the TOX–clinching process using diffuse elements. A computer code baptized SEMA "Static Explicit Method Analysis" is developed to simulate the clinch joining process. The FE code is based on an Updated Lagrangian scheme. The used resolution method is based on an explicit static approach. The integration of the elasto-plastic behavior law is realized with an algorithm of Simo and Taylor. The tools are represented by plane facets.

Keywords: diffuse elements, numerical simulation, clinching, contact, large deformation

Procedia PDF Downloads 337
5518 Regional Review of Outcome of Cervical Smears Reported with Cytological Features of Non Cervical Glandular Neoplasia

Authors: Uma Krishnamoorthy, Vivienne Beavers, Janet Marshall

Abstract:

Introduction: Cervical cytology showing features raising the suspicion of non cervical glandular neoplasia are reported as code 0 under the United Kingdom National Health Service Cervical screening programme ( NHSCSP). As the suspicion is regarding non cervical neoplasia, smear is reported as normal and patient informed that cervical screening result is normal. GP receives copy of results where it states further referral is indicated in small font within text of report. Background: There were several incidents of delayed diagnosis of endometrial cancer in Lancashire which prompted this Northwest Regional review to enable an understanding of underlying pathology outcome of code zero smears to raise awareness and also to review whether further action on wording of smear results was indicated to prevent such delay. Methodology: All Smears reported at the Manchester cytology centre who process cytology for Lancashire population from March 2013 to March 2014 were reviewed and histological diagnosis outcome of women in whom smear was reported as code zero was reviewed retrospectively . Results: Total smears reported by the cytology centre during this period was approximately 109400. Reports issued with result code 0 among this during this time period was 49.Results revealed that among three fourth (37) of women with code zero smear (N=49), evidence of underlying pathology of non cervical origin was confirmed. Of this, 73 % (36) were due to endometrial pathology with 49 % (24) endometrial carcinoma, 12 % (6)polyp, 4 % atypical endometrial hyperplasia (2), 6 % endometrial hyperplasia without atypia (3), and 2 % adenomyosis (1 case) and 2 % ( 1 case) due to ovarian adenocarcinoma. Conclusion: This review demonstrated that more than half (51 %) of women with a code 0 smear report were diagnosed with underlying carcinoma and 75 % had a confirmed underlying pathology contributory to code 0 smear findings. Recommendations and Action Plan: A local rapid access referral and management pathway for this group of women was implemented as a result of this in our unit. The findings and Pathway were shared with other regional units served by the cytology centre through the Pan Lancashire cervical screening board and through the Cytology centre. Locally, the smear report wording was updated to include a rubber stamp/ print in "Red Bold letters" stating that " URGENT REFERRAL TO GYNAECOLOGY IS INDICATED". Findings were also shared through the Pan Lancashire board with National cervical screening programme board, and revisions to wording of code zero smear reports to highlight the need for Urgent referral has now been agreed at National level to be implemented.

Keywords: code zero smears, endometrial cancer, non cervical glandular neoplasia, ovarian cancer

Procedia PDF Downloads 258
5517 Preliminary Analysis for Oil and Gas Geological Characteristics and Exploration Prospects of Doseo Basin in Central Africa

Authors: Haiqiang Song, Huiqing Liu

Abstract:

The Doseo basin in Chad, Central Africa is one of the most important oil and gas blocks in the world. However, the low degree of oil and gas exploration and the lack of relevant geological data restrict the understanding and resource evaluation of the basin. To further develop the Doseo basin efficiently, it is urgent to deeply analyze the source rock characteristics and hydrocarbon generation potential of the Doseo basin. Based on seismic and drilling data in recent years, this paper systematically evaluates the geochemical characteristics of source rocks and their generated oils in Doseo Basin, explores the development, distribution, and evolution characteristics of source rocks, and evaluates the exploration potential of Doseo Basin according to the hydrocarbon enrichment law. The results show that the Lower Cretaceous Baliemian and Apudian source rocks in Doseo Basin are well developed, with high organic matter abundance (average TOC≥3%) and good organic matter types (type I~II), which are the main development layers of source rocks, but the organic matter maturity is generally low (Ro of the drilled source rocks is mainly between 0.4%~0.8%). The planar structure also shows that the main hydrocarbon accumulation mode in Doseo sag is the forward tectonic reservoirs such as near source anticlines and faulted noses. Finally, it is estimated that the accumulative resources of the main source rocks in the Doseo Basin are about 4.33× 108T in Apudite and Balim terrace layers. The results of this study will help guide the next step of oil and gas exploration, which is expected to drive the next step of oil and gas development.

Keywords: Doseo basin, lower cretaceous, source rock characteristics, developmental characteristics, hydrocarbon generation potential

Procedia PDF Downloads 74
5516 A Framework for Secure Information Flow Analysis in Web Applications

Authors: Ralph Adaimy, Wassim El-Hajj, Ghassen Ben Brahim, Hazem Hajj, Haidar Safa

Abstract:

Huge amounts of data and personal information are being sent to and retrieved from web applications on daily basis. Every application has its own confidentiality and integrity policies. Violating these policies can have broad negative impact on the involved company’s financial status, while enforcing them is very hard even for the developers with good security background. In this paper, we propose a framework that enforces security-by-construction in web applications. Minimal developer effort is required, in a sense that the developer only needs to annotate database attributes by a security class. The web application code is then converted into an intermediary representation, called Extended Program Dependence Graph (EPDG). Using the EPDG, the provided annotations are propagated to the application code and run against generic security enforcement rules that were carefully designed to detect insecure information flows as early as they occur. As a result, any violation in the data’s confidentiality or integrity policies is reported. As a proof of concept, two PHP web applications, Hotel Reservation and Auction, were used for testing and validation. The proposed system was able to catch all the existing insecure information flows at their source. Moreover and to highlight the simplicity of the suggested approaches vs. existing approaches, two professional web developers assessed the annotation tasks needed in the presented case studies and provided a very positive feedback on the simplicity of the annotation task.

Keywords: web applications security, secure information flow, program dependence graph, database annotation

Procedia PDF Downloads 439
5515 Natural Convection of a Nanofluid in a Conical Container

Authors: Brahim Mahfoud, Ali Bendjaghlouli

Abstract:

Natural convection is simulated in a truncated cone filled with nanofluid. Inclined and top walls have constant temperature where the heat source is located on the bottom wall of the conical container which is thermally insulated. A finite volume approach is used to solve the governing equations using the SIMPLE algorithm for different parameters such as Rayleigh number, inclination angle of inclined walls of the enclosure and heat source length. The results showed an enhancement in cooling system by using a nanofluid, when conduction regime is assisted. The inclination angle of inclined sidewall and heat source length affect the heat transfer rate and the maximum temperature.

Keywords: heat source, truncated cone, nanofluid, natural convection

Procedia PDF Downloads 339
5514 Steady State Modeling and Simulation of an Industrial Steam Boiler

Authors: Amina Lyria Deghal Cheridi, Abla Chaker, Ahcene Loubar

Abstract:

Relap5 system code is one among powerful tools, which is used in the area of design and safety evaluation. This work aims to simulate the behavior of a radiant steam boiler at the steady-state conditions using Relap5 code system. To perform this study, a detailed Relap5 model is built including all the parts of the steam boiler. The control and regulation systems are also considered. To reproduce the most important parameters and phenomena with an acceptable accuracy and fidelity, a strong qualification work is undertaken concerning the facility nodalization. It consists of making a comparison between the code results and the plant available data in steady-state operation mode. Therefore, the model qualification results at the steady-state are in good agreement with the steam boiler experimental data. The steam boiler Relap5 model has proved satisfactory; and the model was capable of predicting the main thermal-hydraulic steady-state conditions of the steam boiler.

Keywords: industrial steam boiler, model qualification, natural circulation, relap5/mod3.2, steady state simulation

Procedia PDF Downloads 244
5513 Recommended Practice for Experimental Evaluation of the Seepage Sensitivity Damage of Coalbed Methane Reservoirs

Authors: Hao Liu, Lihui Zheng, Chinedu J. Okere, Chao Wang, Xiangchun Wang, Peng Zhang

Abstract:

The coalbed methane (CBM) extraction industry (an unconventional energy source) is yet to promulgated an established standard code of practice for the experimental evaluation of sensitivity damage of coal samples. The existing experimental process of previous researches mainly followed the industry standard for conventional oil and gas reservoirs (CIS). However, the existing evaluation method ignores certain critical differences between CBM reservoirs and conventional reservoirs, which could inevitably result in an inaccurate evaluation of sensitivity damage and, eventually, poor decisions regarding the formulation of formation damage prevention measures. In this study, we propose improved experimental guidelines for evaluating seepage sensitivity damage of CBM reservoirs by leveraging on the shortcomings of the existing methods. The proposed method was established via a theoretical analysis of the main drawbacks of the existing methods and validated through comparative experiments. The results show that the proposed evaluation technique provided reliable experimental results that can better reflect actual reservoir conditions and correctly guide future development of CBM reservoirs. This study is pioneering the research on the optimization of experimental parameters for efficient exploration and development of CBM reservoirs.

Keywords: coalbed methane, formation damage, permeability, unconventional energy source

Procedia PDF Downloads 101
5512 Developing a Framework for Open Source Software Adoption in a Higher Education Institution in Uganda. A case of Kyambogo University

Authors: Kafeero Frank

Abstract:

This study aimed at developing a frame work for open source software adoption in an institution of higher learning in Uganda, with the case of KIU as a study area. There were mainly four research questions based on; individual staff interaction with open source software forum, perceived FOSS characteristics, organizational characteristics and external characteristics as factors that affect open source software adoption. The researcher used causal-correlation research design to study effects of these variables on open source software adoption. A quantitative approach was used in this study with self-administered questionnaire on a purposively and randomly sampled sample of university ICT staff. Resultant data was analyzed using means, correlation coefficients and multivariate multiple regression analysis as statistical tools. The study reveals that individual staff interaction with open source software forum and perceived FOSS characteristics were the primary factors that significantly affect FOSS adoption while organizational and external factors were secondary with no significant effect but significant correlation to open source software adoption. It was concluded that for effective open source software adoption to occur there must be more effort on primary factors with subsequent reinforcement of secondary factors to fulfill the primary factors and adoption of open source software. Lastly recommendations were made in line with conclusions for coming up with Kyambogo University frame work for open source software adoption in institutions of higher learning. Areas of further research recommended include; Stakeholders’ analysis of open source software adoption in Uganda; Challenges and way forward. Evaluation of Kyambogo University frame work for open source software adoption in institutions of higher learning. Framework development for cloud computing adoption in Ugandan universities. Framework for FOSS development in Uganda IT industry

Keywords: open source software., organisational characteristics, external characteristics, cloud computing adoption

Procedia PDF Downloads 36
5511 Growth of Multi-Layered Graphene Using Organic Solvent-PMMA Film as the Carbon Source under Low Temperature Conditions

Authors: Alaa Y. Ali, Natalie P. Holmes, John Holdsworth, Warwick Belcher, Paul Dastoor, Xiaojing Zhou

Abstract:

Multi-layered graphene has been produced under low temperature chemical vapour deposition (CVD) growth conditions by utilizing an organic solvent and polymer film source. Poly(methylmethacrylate) (PMMA) was dissolved in chlorobenzene solvent and used as a drop-cast film carbon source on a quartz slide. A source temperature (Tsource) of 180 °C provided sufficient carbon to grow graphene, as identified by Raman spectroscopy, on clean copper foil catalytic surfaces.  Systematic variation of hydrogen gas (H2) flow rate from 25 standard cubic centimeters per minute (sccm) to 100 sccm and CVD temperature (Tgrowth) from 400 to 800 °C, yielded graphene films of varying quality as characterized by Raman spectroscopy. The optimal graphene growth parameters were found to occur with a hydrogen flow rate of 75 sccm sweeping the 180 °C source carbon past the Cu foil at 600 °C for 1 min. The deposition at 600 °C with a H2 flow rate of 75 sccm yielded a 2D band peak with ~53.4 cm-1 FWHM and a relative intensity ratio of the G to 2D bands (IG/I2D) of 0.21. This recipe fabricated a few layers of good quality graphene.

Keywords: graphene, chemical vapor deposition, carbon source, low temperature growth

Procedia PDF Downloads 136
5510 ROOP: Translating Sequential Code Fragments to Distributed Code Fragments Using Deep Reinforcement Learning

Authors: Arun Sanjel, Greg Speegle

Abstract:

Every second, massive amounts of data are generated, and Data Intensive Scalable Computing (DISC) frameworks have evolved into effective tools for analyzing such massive amounts of data. Since the underlying architecture of these distributed computing platforms is often new to users, building a DISC application can often be time-consuming and prone to errors. The automated conversion of a sequential program to a DISC program will consequently significantly improve productivity. However, synthesizing a user’s intended program from an input specification is complex, with several important applications, such as distributed program synthesizing and code refactoring. Existing works such as Tyro and Casper rely entirely on deductive synthesis techniques or similar program synthesis approaches. Our approach is to develop a data-driven synthesis technique to identify sequential components and translate them to equivalent distributed operations. We emphasize using reinforcement learning and unit testing as feedback mechanisms to achieve our objectives.

Keywords: program synthesis, distributed computing, reinforcement learning, unit testing, DISC

Procedia PDF Downloads 72
5509 Analyzing Oil Seeps Manifestations and Petroleum Impregnation in Northwestern Tunisia From Aliphatic Biomarkers and Statistical Data

Authors: Sawsen Jarray, Tahani Hallek, Mabrouk Montacer

Abstract:

The tectonically damaged terrain in Tunisia's Northwest is seen in the country's numerous oil leaks. Finding a genetic link between these oil seeps and the area's putative source rocks is the goal of this investigation. Here, we use aliphatic biomarkers assessed by GC-MS to describe the organic geochemical data of 18 oil seeps samples and 4 source rocks (M'Cherga, Fahdene, Bahloul, and BouDabbous). In order to establish correlations between oil and oil and oil and source rock, terpanes, hopanes, and steranes biomarkers were identified. The source rocks under study were deposited in a marine environment and were suboxic, with minor signs of continental input for the M'Cherga Formation. There is no connection between the Fahdene and Bahloul source rocks and the udied oil seeps. According to the biomarkers C27 18-22,29,30trisnorneohopane (Ts) and C27 17-22,29,30-trisnorhopane (Tm), these source rocks are mature and have reached the oil window. Regarding oil seeps, geochemical data indicate that, with the exception of four samples that showed some continental markings, the bulk of samples were deposited in an open marine environment. These most recent samples from oil seeps have a unique lithology (marl) that distinguishes them from the others (carbonate). There are two classes of oil seeps, according to statistical analysis of relationships between oil and oil and oil and source rocks. The first comprised samples that showed a positive connection with carbonate-lithological and marine-derived BouDabbous black shales. The second is a result of M'Cherga source rock and is made up of oil seeps with remnants of the terrestrial environment and a lithology with a marl trend. The Fahdene and Bahloul source rocks have no connection to the observed oil seeps. There are two different types of hydrocarbon spills depending on their link to tectonic deformations (oil seeps) and outcropping mature source rocks (oil impregnations), in addition to the existence of two generations of hydrocarbon spills in Northwest Tunisia (Lower Cretaceous/Ypresian).

Keywords: petroleum seeps, source rocks, biomarkers, statistic, Northern Tunisia

Procedia PDF Downloads 35
5508 PointNetLK-OBB: A Point Cloud Registration Algorithm with High Accuracy

Authors: Wenhao Lan, Ning Li, Qiang Tong

Abstract:

To improve the registration accuracy of a source point cloud and template point cloud when the initial relative deflection angle is too large, a PointNetLK algorithm combined with an oriented bounding box (PointNetLK-OBB) is proposed. In this algorithm, the OBB of a 3D point cloud is used to represent the macro feature of source and template point clouds. Under the guidance of the iterative closest point algorithm, the OBB of the source and template point clouds is aligned, and a mirror symmetry effect is produced between them. According to the fitting degree of the source and template point clouds, the mirror symmetry plane is detected, and the optimal rotation and translation of the source point cloud is obtained to complete the 3D point cloud registration task. To verify the effectiveness of the proposed algorithm, a comparative experiment was performed using the publicly available ModelNet40 dataset. The experimental results demonstrate that, compared with PointNetLK, PointNetLK-OBB improves the registration accuracy of the source and template point clouds when the initial relative deflection angle is too large, and the sensitivity of the initial relative position between the source point cloud and template point cloud is reduced. The primary contribution of this paper is the use of PointNetLK to avoid the non-convex problem of traditional point cloud registration and leveraging the regularity of the OBB to avoid the local optimization problem in the PointNetLK context.

Keywords: mirror symmetry, oriented bounding box, point cloud registration, PointNetLK-OBB

Procedia PDF Downloads 110
5507 Pallet Tracking and Cost Optimization of the Flow of Goods in Logistics Operations by Serial Shipping Container Code

Authors: Dominika Crnjac Milic, Martina Martinovic, Vladimir Simovic

Abstract:

The case study method in this paper shows the implementation of Information Technology (IT) and the Serial Shipping Container Code (SSCC) in a Croatian company that deals with logistics operations and provides logistics services in the cold chain segment. This company is aware of the sensitivity of the goods entrusted to them by the user of the service, as well as of the importance of speed and accuracy in providing logistics services. To that end, it has implemented and used the latest IT to ensure the highest standard of high-quality logistics services to its customers. Looking for efficiency and optimization of supply chain management, while maintaining a high level of quality of the products that are sold, today's users of outsourced logistics services are open to the implementation of new IT products that ultimately deliver savings. By analysing the positive results and the difficulties that arise when using this technology, we aim to provide an insight into the potential of this approach of the logistics service provider.

Keywords: logistics operations, serial shipping container code, information technology, cost optimization

Procedia PDF Downloads 338
5506 Model Driven Architecture Methodologies: A Review

Authors: Arslan Murtaza

Abstract:

Model Driven Architecture (MDA) is technique presented by OMG (Object Management Group) for software development in which different models are proposed and converted them into code. The main plan is to identify task by using PIM (Platform Independent Model) and transform it into PSM (Platform Specific Model) and then converted into code. In this review paper describes some challenges and issues that are faced in MDA, type and transformation of models (e.g. CIM, PIM and PSM), and evaluation of MDA-based methodologies.

Keywords: OMG, model driven rrchitecture (MDA), computation independent model (CIM), platform independent model (PIM), platform specific model(PSM), MDA-based methodologies

Procedia PDF Downloads 414
5505 Comparison of Polyphonic Profile of a Berry from Two Different Sources, Using an Optimized Extraction Method

Authors: G. Torabian, A. Fathi, P. Valtchev, F. Dehghani

Abstract:

The superior polyphenol content of Sambucus nigra berries has high health potentials for the production of nutraceutical products. Numerous factors influence the polyphenol content of the final products including the berries’ source and the subsequent processing production steps. The aim of this study is to compare the polyphenol content of berries from two different sources and also to optimise the polyphenol extraction process from elderberries. Berries from source B obtained more acceptable physical properties than source A; a single berry from source B was double in size and weight (both wet and dry weight) compared with a source A berry. Despite the appropriate physical characteristics of source B berries, their polyphenolic profile was inferior; as source A berries had 2.3 fold higher total anthocyanin content, and nearly two times greater total phenolic content and total flavonoid content compared to source B. Moreover, the result of this study showed that almost 50 percent of the phenolic content of berries are entrapped within their skin and pulp that potentially cannot be extracted by press juicing. To address this challenge and to increase the total polyphenol yield of the extract, we used cold-shock blade grinding method to break the cell walls. The result of this study showed that using cultivars with higher phenolic content as well as using the whole fruit including juice, skin and pulp can increase polyphenol yield significantly; and thus, may boost the potential of using elderberries as therapeutic products.

Keywords: different sources, elderberry, grinding, juicing, polyphenols

Procedia PDF Downloads 264
5504 2D and 3D Unsteady Simulation of the Heat Transfer in the Sample during Heat Treatment by Moving Heat Source

Authors: Zdeněk Veselý, Milan Honner, Jiří Mach

Abstract:

The aim of the performed work is to establish the 2D and 3D model of direct unsteady task of sample heat treatment by moving source employing computer model on the basis of finite element method. The complex boundary condition on heat loaded sample surface is the essential feature of the task. Computer model describes heat treatment of the sample during heat source movement over the sample surface. It is started from the 2D task of sample cross section as a basic model. Possibilities of extension from 2D to 3D task are discussed. The effect of the addition of third model dimension on the temperature distribution in the sample is showed. Comparison of various model parameters on the sample temperatures is observed. Influence of heat source motion on the depth of material heat treatment is shown for several velocities of the movement. Presented computer model is prepared for the utilization in laser treatment of machine parts.

Keywords: computer simulation, unsteady model, heat treatment, complex boundary condition, moving heat source

Procedia PDF Downloads 360
5503 Implementation of an Open Source ERP for SMEs in the Automotive Sector in Peru: A Case Study

Authors: Gerson E. Cornejo, Luis A. Gamarra, David S. Mauricio

Abstract:

The Enterprise Resource Planning Systems (ERP) allows the integration of all the business processes of the functional areas of the companies, in order to automate and standardize the processes, obtain accurate information and improve decision making in time real. In Peru, 79% of medium and small companies (SMEs) do not use any management software, this is because it is believed that ERPs are expensive, complex and difficult to implement. However, for more than 20 years there have been Open Source ERPs, which are more accessible and have the same benefit as proprietary ERPs, but there is little information on the implementation process. In this work is made a case of study, in order to show the implementation process of an Open Source ERP, Odoo, based on the ASAP methodology (Accelerated SAP) and applied to a company of corrective and preventive maintenance services of vehicles. The ERP allowed the SME to standardize its business processes, increase its productivity, reducing up to 40% certain processes. The study of this case shows that it is feasible and profitable to implement an Open Source ERP in SMEs in the Automotive Sector of Peru. In addition, it is shown that the ASAP methodology is adequate to carry out Open Source ERPs implementation projects.

Keywords: ASAP, automotive sector, ERP implementation, open source

Procedia PDF Downloads 301
5502 Evaluation of Response Modification Factors in Moment Resisting Frame Buildings Considering Soil Structure Interaction

Authors: K. Farheen, A. Munir

Abstract:

Seismic response of the multi-storey buildings is created by the interaction of both the structure and underlying soil medium. The seismic design philosophy is incorporated using response modification factor 'R'. Current code based values of 'R' factor does not reflect the SSI problem as it is based on fixed base condition. In this study, the modified values of 'R' factor for moment resisting frame (MRF) considering SSI are evaluated. The response of structure with and without SSI has been compared using equivalent linear static and nonlinear static pushover analyses for 10-storied moment resisting frame building. The building is located in seismic zone 2B situated on different soils with shear wave velocity (Vₛ) of 300m/sec (SD) and 1200m/s (SB). Code based 'R' factor value for building frame system has been taken as 5.5. Soil medium is modelled using identical but mutually independent horizontal and vertical springs. It was found that the modified 'R' factor values have been decreased by 47% and 43% for soil SD and SB respectively as compared to that of code based 'R' factor.

Keywords: buildings, SSI, shear wave velocity, R factor

Procedia PDF Downloads 178
5501 GA3C for Anomalous Radiation Source Detection

Authors: Chia-Yi Liu, Bo-Bin Xiao, Wen-Bin Lin, Hsiang-Ning Wu, Liang-Hsun Huang

Abstract:

In order to reduce the risk of radiation damage that personnel may suffer during operations in the radiation environment, the use of automated guided vehicles to assist or replace on-site personnel in the radiation environment has become a key technology and has become an important trend. In this paper, we demonstrate our proof of concept for autonomous self-learning radiation source searcher in an unknown environment without a map. The research uses GPU version of Asynchronous Advantage Actor-Critic network (GA3C) of deep reinforcement learning to search for radiation sources. The searcher network, based on GA3C architecture, has self-directed learned and improved how search the anomalous radiation source by training 1 million episodes under three simulation environments. In each episode of training, the radiation source position, the radiation source intensity, starting position, are all set randomly in one simulation environment. The input for searcher network is the fused data from a 2D laser scanner and a RGB-D camera as well as the value of the radiation detector. The output actions are the linear and angular velocities. The searcher network is trained in a simulation environment to accelerate the learning process. The well-performance searcher network is deployed to the real unmanned vehicle, Dashgo E2, which mounts LIDAR of YDLIDAR G4, RGB-D camera of Intel D455, and radiation detector made by Institute of Nuclear Energy Research. In the field experiment, the unmanned vehicle is enable to search out the radiation source of the 18.5MBq Na-22 by itself and avoid obstacles simultaneously without human interference.

Keywords: deep reinforcement learning, GA3C, source searching, source detection

Procedia PDF Downloads 82
5500 Reduction of Multiple User Interference for Optical CDMA Systems Using Successive Interference Cancellation Scheme

Authors: Tawfig Eltaif, Hesham A. Bakarman, N. Alsowaidi, M. R. Mokhtar, Malek Harbawi

Abstract:

In Commonly, it is primary problem that there is multiple user interference (MUI) noise resulting from the overlapping among the users in optical code-division multiple access (OCDMA) system. In this article, we aim to mitigate this problem by studying an interference cancellation scheme called successive interference cancellation (SIC) scheme. This scheme will be tested on two different detection schemes, spectral amplitude coding (SAC) and direct detection systems (DS), using partial modified prime (PMP) as the signature codes. It was found that SIC scheme based on both SAC and DS methods had a potential to suppress the intensity noise, that is to say, it can mitigate MUI noise. Furthermore, SIC/DS scheme showed much lower bit error rate (BER) performance relative to SIC/SAC scheme for different magnitude of effective power. Hence, many more users can be supported by SIC/DS receiver system.

Keywords: optical code-division multiple access (OCDMA), successive interference cancellation (SIC), multiple user interference (MUI), spectral amplitude coding (SAC), partial modified prime code (PMP)

Procedia PDF Downloads 490
5499 Cognitive Function During the First Two Hours of Spravato Administration in Patients with Major Depressive Disorder

Authors: Jocelyn Li, Xiangyang Li

Abstract:

We have employed THINC-it® to study the acute effects of Spravato on the cognitive function of patients with severe major depression disorder (MDD). The scores of the four tasks (Spotter, Symbol Check, Code Breaker, Trails) found in THINC-it® were used to measure cognitive function throughout treatment. The patients who participated in this study have tried more than 3 antidepressants without significant improvement before they began Spravato treatment. All patients received 3 doses of 28 mg Spravato 5 minutes apart (84 mg total per treatment) during this study with THINC-it®. The data were collected before the first Spravato administration (T0), 1 hour after the first Spravato administration (T1), and 2 hours after the first Spravato administration (T2) during each treatment. The following data were from 13 patients, with a total of 226 trials in a 2-3 month period. Spravato at 84 mg reduced the scores of Trails, Code Breaker, Symbol Check, and Spotter at T1 by 10-20% in all patients with one exception for a minority of patients in Spotter. At T2, the scores of Trails, Symbol Check, and Spotter were back to 97% of T0 while the score of Code Breaker was back to 92%. Interestingly, we found that the score of Spotter was consistently increased by 17% at T1 in the same 30% of patients in each treatment. We called this change reverse response while the pattern of the other patients, a decline (T1) and then recovery (T2), was called non-reverse response. We also compared the scores at T0 between the first visit and the fifth visit. The T0 scores of all four tasks were improved at visit 5 when compared to visit 1. The scores of Trails, Code Breaker, and Symbol Check at T0 were increased by 14%, 33%, and 14% respectively at visit 5. The score of Code Breaker, which had two trends, improved by 9% in reverse response patients compared to a 27% improvement in non-reverse response patients. To our knowledge, this is the first study done on the impact of Spravato on cognitive function change in major depression patients at this time frame. Whether we can predict future responses to Spravato with THINC-it® merits further study.

Keywords: Spravato, THINC-it, major depressive disorder, cognitive function

Procedia PDF Downloads 86
5498 Surface to the Deeper: A Universal Entity Alignment Approach Focusing on Surface Information

Authors: Zheng Baichuan, Li Shenghui, Li Bingqian, Zhang Ning, Chen Kai

Abstract:

Entity alignment (EA) tasks in knowledge graphs often play a pivotal role in the integration of knowledge graphs, where structural differences often exist between the source and target graphs, such as the presence or absence of attribute information and the types of attribute information (text, timestamps, images, etc.). However, most current research efforts are focused on improving alignment accuracy, often along with an increased reliance on specific structures -a dependency that inevitably diminishes their practical value and causes difficulties when facing knowledge graph alignment tasks with varying structures. Therefore, we propose a universal knowledge graph alignment approach that only utilizes the common basic structures shared by knowledge graphs. We have demonstrated through experiments that our method achieves state-of-the-art performance in fair comparisons.

Keywords: knowledge graph, entity alignment, transformer, deep learning

Procedia PDF Downloads 0