Search results for: computer tasks
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3743

Search results for: computer tasks

2423 Outdoor Anomaly Detection with a Spectroscopic Line Detector

Authors: O. J. G. Somsen

Abstract:

One of the tasks of optical surveillance is to detect anomalies in large amounts of image data. However, if the size of the anomaly is very small, limited information is available to distinguish it from the surrounding environment. Spectral detection provides a useful source of additional information and may help to detect anomalies with a size of a few pixels or less. Unfortunately, spectral cameras are expensive because of the difficulty of separating two spatial in addition to one spectral dimension. We investigate the possibility of modifying a simpler spectral line detector for outdoor detection. This may be especially useful if the area of interest forms a line, such as the horizon. We use a monochrome CCD that also enables detection into the near infrared. A simple camera is attached to the setup to determine which part of the environment is spectrally imaged. Our preliminary results indicate that sensitive detection of very small targets is indeed possible. Spectra could be taken from the various targets by averaging columns in the line image. By imaging a set of lines of various width we found narrow lines that could not be seen in the color image but remained visible in the spectral line image. A simultaneous analysis of the entire spectra can produce better results than visual inspection of the line spectral image. We are presently developing calibration targets for spatial and spectral focusing and alignment with the spatial camera. This will present improved results and more use in outdoor application

Keywords: anomaly detection, spectroscopic line imaging, image analysis, outdoor detection

Procedia PDF Downloads 481
2422 Electronic Data Interchange (EDI) in the Supply Chain: Impact on Customer Satisfaction

Authors: Hicham Amine, Abdelouahab Mesnaoui

Abstract:

Electronic data interchange EDI is the computer-to-computer exchange of structured business information. This information typically takes the form of standardized electronic business documents, such as invoices, purchase orders, bills of lading, and so on. The purpose of this study is to identify the impact EDI might have on supply chain and typically on customer satisfaction keeping in mind the constraints the organization might face. This study included 139 subject matter experts (SMEs) who participated by responding to a survey that was distributed. 85% responded that they are extremely for the implementation while 10% were neutral and 5% were against the implementation. From the quality assurance department, we have got 75% from the clients agreed to move on with the change whereas 10% stayed neutral and finally 15% were against the change. From the legal department where 80% of the answers were for the implementation and 10% of the participants stayed neutral whereas the last 10% were against it. The survey consisted of 40% male and 60% female (sex-ratio (F/M=1,5), who had chosen to participate. Our survey also contained 3 categories in terms of technical background where 80% are from technical background and 15% were from nontechnical background and 5% had some average technical background. This study examines the impact of EDI on customer satisfaction which is the primary hypothesis and justifies the importance of the implementation which enhances the customer satisfaction.

Keywords: electronic data interchange, supply chain, subject matter experts, customer satisfaction

Procedia PDF Downloads 340
2421 Corporate Governance and Business Ethical Values in Organisation: A Study of Unilag Holdings

Authors: Aribisala Oluwadamilare Olufolarin

Abstract:

The objective of this research was to examine how corporate governance and ethical business values impact both the performance of the organization and its employees, as it is essential for any organization to uphold good ethics and corporate governance. The study was conducted at Unilag Holdings Limited (UniHOLDs) to demonstrate that organizations may experience losses if they do not have proper corporate governance and business ethical values in place. The employees' perception of corporate governance and ethics is crucial for the organization. The research indicated a connection between corporate governance and business ethics values, and therefore, correlation analysis was utilized, making it statistically reliable. The results of the test show a strong positive correlation (r=.812, N=94, P<.01) between corporate governance and business ethical values. A questionnaire was distributed to employees at Unilag Holdings Limited (UniHOLDs), with 94 out of 130 completed and returned. The findings indicate that ethical values contribute to employee productivity, and productive employees have a beneficial impact on the organization's performance. Additionally, the study revealed that employees tend to adhere to rules regardless of their ethical nature. To address this, the organization should ensure that top-level managers do not assign unethical tasks to their subordinates. The study recommends that the organization should consistently practice corporate governance and business ethics. The company needs to make sure that its stakeholders continue to support its way of doing things.

Keywords: business ethics, business ethical values, corporate governance and organization, corporate governance

Procedia PDF Downloads 12
2420 Performance Comparison of Thread-Based and Event-Based Web Servers

Authors: Aikaterini Kentroti, Theodore H. Kaskalis

Abstract:

Today, web servers are expected to serve thousands of client requests concurrently within stringent response time limits. In this paper, we evaluate experimentally and compare the performance as well as the resource utilization of popular web servers, which differ in their approach to handle concurrency. More specifically, Central Processing Unit (CPU)- and I/O intensive tests were conducted against the thread-based Apache and Go as well as the event-based Nginx and Node.js under increasing concurrent load. The tests involved concurrent users requesting a term of the Fibonacci sequence (the 10th, 20th, 30th) and the content of a table from the database. The results show that Go achieved the best performance in all benchmark tests. For example, Go reached two times higher throughput than Node.js and five times higher than Apache and Nginx in the 20th Fibonacci term test. In addition, Go had the smallest memory footprint and demonstrated the most efficient resource utilization, in terms of CPU usage. Instead, Node.js had by far the largest memory footprint, consuming up to 90% more memory than Nginx and Apache. Regarding the performance of Apache and Nginx, our findings indicate that Hypertext Preprocessor (PHP) becomes a bottleneck when the servers are requested to respond by performing CPU-intensive tasks under increasing concurrent load.

Keywords: apache, Go, Nginx, node.js, web server benchmarking

Procedia PDF Downloads 97
2419 Enhancing Disaster Response Capabilities in Asia-Pacific: An Explorative Study Applied to Decision Support Tools for Logistics Network Design

Authors: Giuseppe Timperio, Robert de Souza

Abstract:

Logistics operations in the context of disaster response are characterized by a high degree of complexity due to the combined effect of a large number of stakeholders involved, time pressure, uncertainties at various levels, massive deployment of goods and personnel, and gigantic financial flow to be managed. It also involves several autonomous parties such as government agencies, militaries, NGOs, UN agencies, private sector to name few, to have a highly collaborative approach especially in the critical phase of the immediate response. This is particularly true in the context of L3 emergencies that are the most severe, large-scale humanitarian crises. Decision-making processes in disaster management are thus extremely difficult due to the presence of multiple decision-makers involved, and the complexity of the tasks being tackled. Hence, in this paper, we look at applying ICT based solutions to enable a speedy and effective decision making in the golden window of humanitarian operations. A high-level view of ICT based solutions in the context of logistics operations for humanitarian response in Southeast Asia is presented, and their viability in a real-life case about logistics network design is explored.

Keywords: decision support, disaster preparedness, humanitarian logistics, network design

Procedia PDF Downloads 168
2418 Placement of English Lexical Stress by Arabic-Speaking EFL Learners: How Computer-Generated Spectrographic Representations of Correct Pronunciations Can Provide a Visual Aid to Learners

Authors: Rami Al-Sadi

Abstract:

The assignment of lexical stress in English to its correct syllable is an enormous challenge to EFL learners, especially if their first language (L1) phonology is very different from English phonology. Arabic-speaking EFL learners not only stumble very frequently when it comes to placing the lexical stress in a given word, but they also seem to relegate the role of lexical stress as unimportant, mainly because in Arabic, unlike in English, lexical stress is not phonemic. This study aims at exploring the possible benefits of utilizing spectrographic representations of English words correctly pronounced, for the purpose of finding out how these spectrograms can provide a visual aid to the learners and help them rectify their stress placement errors as they see in real time spectrograms of the correct pronunciations juxtaposed on a computer screen with spectrograms of their own pronunciations for easy comparison. The study involved 120 students from the English Department at Prince Sattam bin Abdulaziz University in Saudi Arabia. 60 participants were taught the English lexical stress rules and also received spectrographic guidance on pronunciation; the other 60 received only verbal instruction on the stress rules and verbal feedback on their pronunciations. Statistical results showed that when the learners had the opportunity to ‘see’ their pronunciation mistakes, they were three times more likely to rectify their placement of lexical stress.

Keywords: Arabic-speaking EFL learners, lexical stress, pronunciation, spectrographic representation, stress placement

Procedia PDF Downloads 123
2417 Performants: Making the Organization of Concerts Easier

Authors: Ioannis Andrianakis, Panagiotis Panagiotopoulos, Kyriakos Chatzidimitriou, Dimitrios Tampakis, Manolis Falelakis

Abstract:

Live music, whether performed in organized venues, restaurants, hotels or any other spots, creates value chains that support and develop local economies and tourism development. In this paper, we describe PerformAnts, a platform that increases the mobility of musicians and their accessibility to remotely located venues by rationalizing the cost of live acts. By analyzing the event history and taking into account their potential availability, the platform provides bespoke recommendations to both bands and venues while also facilitating the organization of tours and helping rationalize transportation expenses by realizing an innovative mechanism called “chain booking”. Moreover, the platform provides an environment where complicated tasks such as technical and financial negotiations, concert promotion or copyrights are easily manipulated by users using best practices. The proposed solution provides important benefits to the whole spectrum of small/medium size concert organizers, as the complexity and the cost of the production are rationalized. The environment is also very beneficial for local talent, musicians that are very mobile, venues located away from large urban areas or in touristic destinations, and managers who will be in a position to coordinate a larger number of musicians without extra effort.

Keywords: machine learning, music industry, creative industries, web applications

Procedia PDF Downloads 97
2416 Predicting the Diagnosis of Alzheimer’s Disease: Development and Validation of Machine Learning Models

Authors: Jay L. Fu

Abstract:

Patients with Alzheimer's disease progressively lose their memory and thinking skills and, eventually, the ability to carry out simple daily tasks. The disease is irreversible, but early detection and treatment can slow down the disease progression. In this research, publicly available MRI data and demographic data from 373 MRI imaging sessions were utilized to build models to predict dementia. Various machine learning models, including logistic regression, k-nearest neighbor, support vector machine, random forest, and neural network, were developed. Data were divided into training and testing sets, where training sets were used to build the predictive model, and testing sets were used to assess the accuracy of prediction. Key risk factors were identified, and various models were compared to come forward with the best prediction model. Among these models, the random forest model appeared to be the best model with an accuracy of 90.34%. MMSE, nWBV, and gender were the three most important contributing factors to the detection of Alzheimer’s. Among all the models used, the percent in which at least 4 of the 5 models shared the same diagnosis for a testing input was 90.42%. These machine learning models allow early detection of Alzheimer’s with good accuracy, which ultimately leads to early treatment of these patients.

Keywords: Alzheimer's disease, clinical diagnosis, magnetic resonance imaging, machine learning prediction

Procedia PDF Downloads 143
2415 Artificial Intelligence Impact on the Australian Government Public Sector

Authors: Jessica Ho

Abstract:

AI has helped government, businesses and industries transform the way they do things. AI is used in automating tasks to improve decision-making and efficiency. AI is embedded in sensors and used in automation to help save time and eliminate human errors in repetitive tasks. Today, we saw the growth in AI using the collection of vast amounts of data to forecast with greater accuracy, inform decision-making, adapt to changing market conditions and offer more personalised service based on consumer habits and preferences. Government around the world share the opportunity to leverage these disruptive technologies to improve productivity while reducing costs. In addition, these intelligent solutions can also help streamline government processes to deliver more seamless and intuitive user experiences for employees and citizens. This is a critical challenge for NSW Government as we are unable to determine the risk that is brought by the unprecedented pace of adoption of AI solutions in government. Government agencies must ensure that their use of AI complies with relevant laws and regulatory requirements, including those related to data privacy and security. Furthermore, there will always be ethical concerns surrounding the use of AI, such as the potential for bias, intellectual property rights and its impact on job security. Within NSW’s public sector, agencies are already testing AI for crowd control, infrastructure management, fraud compliance, public safety, transport, and police surveillance. Citizens are also attracted to the ease of use and accessibility of AI solutions without requiring specialised technical skills. This increased accessibility also comes with balancing a higher risk and exposure to the health and safety of citizens. On the other side, public agencies struggle with keeping up with this pace while minimising risks, but the low entry cost and open-source nature of generative AI led to a rapid increase in the development of AI powered apps organically – “There is an AI for That” in Government. Other challenges include the fact that there appeared to be no legislative provisions that expressly authorise the NSW Government to use an AI to make decision. On the global stage, there were too many actors in the regulatory space, and a sovereign response is needed to minimise multiplicity and regulatory burden. Therefore, traditional corporate risk and governance framework and regulation and legislation frameworks will need to be evaluated for AI unique challenges due to their rapidly evolving nature, ethical considerations, and heightened regulatory scrutiny impacting the safety of consumers and increased risks for Government. Creating an effective, efficient NSW Government’s governance regime, adapted to the range of different approaches to the applications of AI, is not a mere matter of overcoming technical challenges. Technologies have a wide range of social effects on our surroundings and behaviours. There is compelling evidence to show that Australia's sustained social and economic advancement depends on AI's ability to spur economic growth, boost productivity, and address a wide range of societal and political issues. AI may also inflict significant damage. If such harm is not addressed, the public's confidence in this kind of innovation will be weakened. This paper suggests several AI regulatory approaches for consideration that is forward-looking and agile while simultaneously fostering innovation and human rights. The anticipated outcome is to ensure that NSW Government matches the rising levels of innovation in AI technologies with the appropriate and balanced innovation in AI governance.

Keywords: artificial inteligence, machine learning, rules, governance, government

Procedia PDF Downloads 70
2414 CTHTC: A Convolution-Backed Transformer Architecture for Temporal Knowledge Graph Embedding with Periodicity Recognition

Authors: Xinyuan Chen, Mohd Nizam Husen, Zhongmei Zhou, Gongde Guo, Wei Gao

Abstract:

Temporal Knowledge Graph Completion (TKGC) has attracted increasing attention for its enormous value; however, existing models lack capabilities to capture both local interactions and global dependencies simultaneously with evolutionary dynamics, while the latest achievements in convolutions and Transformers haven't been employed in this area. What’s more, periodic patterns in TKGs haven’t been fully explored either. To this end, a multi-stage hybrid architecture with convolution-backed Transformers is introduced in TKGC tasks for the first time combining the Hawkes process to model evolving event sequences in a continuous-time domain. In addition, the seasonal-trend decomposition is adopted to identify periodic patterns. Experiments on six public datasets are conducted to verify model effectiveness against state-of-the-art (SOTA) methods. An extensive ablation study is carried out accordingly to evaluate architecture variants as well as the contributions of independent components in addition, paving the way for further potential exploitation. Besides complexity analysis, input sensitivity and safety challenges are also thoroughly discussed for comprehensiveness with novel methods.

Keywords: temporal knowledge graph completion, convolution, transformer, Hawkes process, periodicity

Procedia PDF Downloads 78
2413 Paddy/Rice Singulation for Determination of Husking Efficiency and Damage Using Machine Vision

Authors: M. Shaker, S. Minaei, M. H. Khoshtaghaza, A. Banakar, A. Jafari

Abstract:

In this study a system of machine vision and singulation was developed to separate paddy from rice and determine paddy husking and rice breakage percentages. The machine vision system consists of three main components including an imaging chamber, a digital camera, a computer equipped with image processing software. The singulation device consists of a kernel holding surface, a motor with vacuum fan, and a dimmer. For separation of paddy from rice (in the image), it was necessary to set a threshold. Therefore, some images of paddy and rice were sampled and the RGB values of the images were extracted using MATLAB software. Then mean and standard deviation of the data were determined. An Image processing algorithm was developed using MATLAB to determine paddy/rice separation and rice breakage and paddy husking percentages, using blue to red ratio. Tests showed that, a threshold of 0.75 is suitable for separating paddy from rice kernels. Results from the evaluation of the image processing algorithm showed that the accuracies obtained with the algorithm were 98.36% and 91.81% for paddy husking and rice breakage percentage, respectively. Analysis also showed that a suction of 45 mmHg to 50 mmHg yielding 81.3% separation efficiency is appropriate for operation of the kernel singulation system.

Keywords: breakage, computer vision, husking, rice kernel

Procedia PDF Downloads 381
2412 Development of 3D Laser Scanner for Robot Navigation

Authors: Ali Emre Öztürk, Ergun Ercelebi

Abstract:

Autonomous robotic systems needs an equipment like a human eye for their movement. Robotic camera systems, distance sensors and 3D laser scanners have been used in the literature. In this study a 3D laser scanner has been produced for those autonomous robotic systems. In general 3D laser scanners are using 2 dimension laser range finders that are moving on one-axis (1D) to generate the model. In this study, the model has been obtained by a one-dimensional laser range finder that is moving in two –axis (2D) and because of this the laser scanner has been produced cheaper. Furthermore for the laser scanner a motor driver, an embedded system control board has been used and at the same time a user interface card has been used to make the communication between those cards and computer. Due to this laser scanner, the density of the objects, the distance between the objects and the necessary path ways for the robot can be calculated. The data collected by the laser scanner system is converted in to cartesian coordinates to be modeled in AutoCAD program. This study shows also the synchronization between the computer user interface, AutoCAD and the embedded systems. As a result it makes the solution cheaper for such systems. The scanning results are enough for an autonomous robot but the scan cycle time should be developed. This study makes also contribution for further studies between the hardware and software needs since it has a powerful performance and a low cost.

Keywords: 3D laser scanner, embedded system, 1D laser range finder, 3D model

Procedia PDF Downloads 275
2411 Using Indigenous Games to Demystify Probability Theorem in Ghanaian Classrooms: Mathematical Analysis of Ampe

Authors: Peter Akayuure, Michael Johnson Nabie

Abstract:

Similar to many colonized nations in the world, one indelible mark left by colonial masters after Ghana’s independence in 1957 has been the fact that many contexts used to teach statistics and probability concepts are often alien and do not resonate with the social domain of our indigenous Ghanaian child. This has seriously limited the understanding, discoveries, and applications of mathematics for national developments. With the recent curriculum demands of making the Ghanaian child mathematically literate, this qualitative study involved video recordings and mathematical analysis of play sessions of an indigenous girl game called Ampe with the aim to demystify the concepts in probability theorem, which is applied in mathematics related fields of study. The mathematical analysis shows that the game of Ampe, which is widely played by school girls in Ghana, is suitable for learning concepts of the probability theorems. It was also revealed that as a girl game, the use of Ampe provides good lessons to educators, textbook writers, and teachers to rethink about the selection of mathematics tasks and learning contexts that are sensitive to gender. As we undertake to transform teacher education and student learning, the use of indigenous games should be critically revisited.

Keywords: Ampe, mathematical analysis, probability theorem, Ghanaian girl game

Procedia PDF Downloads 370
2410 The Global-Local Dimension in Cognitive Control after Left Lateral Prefrontal Cortex Damage: Evidence from the Non-Verbal Domain

Authors: Eleni Peristeri, Georgia Fotiadou, Ianthi-Maria Tsimpli

Abstract:

The local-global dimension has been studied extensively in healthy controls and preference for globally processed stimuli has been validated in both the visual and auditory modalities. Critically, the local-global dimension has an inherent interference resolution component, a type of cognitive control, and left-prefrontal-cortex-damaged (LPFC) individuals have exhibited inability to override habitual response behaviors in item recognition tasks that involve representational interference. Eight patients with damage in the left PFC (age range: 32;5 to 69;0. Mean age: 54;6 yrs) and twenty age- and education-matched language-unimpaired adults (mean age: 56;7yrs) have participated in the study. Distinct performance patterns were found between the language-unimpaired and the LPFC-damaged group which have mainly stemmed from the latter’s difficulty with inhibiting global stimuli in incongruent trials. Overall, the local-global attentional dimension affects LPFC-damaged individuals with non-fluent aphasia in non-language domains implicating distinct types of inhibitory processes depending on the level of processing.

Keywords: left lateral prefrontal cortex damage (LPFC), local-global non-language attention, representational interference, non-fluent aphasia

Procedia PDF Downloads 470
2409 The Effectiveness of Implementing Interactive Training for Teaching Kazakh Language

Authors: Samal Abzhanova, Saule Mussabekova

Abstract:

Today, a new system of education is being created in Kazakhstan in order to develop the system of education and to satisfy the world class standards. For this purpose, there have been established new requirements and responsibilities to the instructors. Students should not be limited with providing only theoretical knowledge. Also, they should be encouraged to be competitive, to think creatively and critically. Moreover, students should be able to implement these skills into practice. These issues could be resolved through the permanent improvement of teaching methods. Therefore, a specialist who teaches the languages should use up-to-date methods and introduce new technologies. The result of the investigation suggests that an interactive teaching method is one of the new technologies in this field. This paper aims to provide information about implementing new technologies in the process of teaching language. The paper will discuss about necessity of introducing innovative technologies and the techniques of organizing interactive lessons. At the same time, the structure of the interactive lesson, conditions, principles, discussions, small group works and role-playing games will be considered. Interactive methods are carried out with the help of several types of activities, such as working in a team (with two or more group of people), playing situational or role-playing games, working with different sources of information, discussions, presentations, creative works and learning through solving situational tasks and etc.

Keywords: interactive education, interactive methods, system of education, teaching a language

Procedia PDF Downloads 294
2408 Exploring Data Leakage in EEG Based Brain-Computer Interfaces: Overfitting Challenges

Authors: Khalida Douibi, Rodrigo Balp, Solène Le Bars

Abstract:

In the medical field, applications related to human experiments are frequently linked to reduced samples size, which makes the training of machine learning models quite sensitive and therefore not very robust nor generalizable. This is notably the case in Brain-Computer Interface (BCI) studies, where the sample size rarely exceeds 20 subjects or a few number of trials. To address this problem, several resampling approaches are often used during the data preparation phase, which is an overly critical step in a data science analysis process. One of the naive approaches that is usually applied by data scientists consists in the transformation of the entire database before the resampling phase. However, this can cause model’ s performance to be incorrectly estimated when making predictions on unseen data. In this paper, we explored the effect of data leakage observed during our BCI experiments for device control through the real-time classification of SSVEPs (Steady State Visually Evoked Potentials). We also studied potential ways to ensure optimal validation of the classifiers during the calibration phase to avoid overfitting. The results show that the scaling step is crucial for some algorithms, and it should be applied after the resampling phase to avoid data leackage and improve results.

Keywords: data leackage, data science, machine learning, SSVEP, BCI, overfitting

Procedia PDF Downloads 153
2407 Predictive Functional Control with Disturbance Observer for Tendon-Driven Balloon Actuator

Authors: Jun-ya Nagase, Toshiyuki Satoh, Norihiko Saga, Koichi Suzumori

Abstract:

In recent years, Japanese society has been aging, engendering a labour shortage of young workers. Robots are therefore expected to perform tasks such as rehabilitation, nursing elderly people, and day-to-day work support for elderly people. The pneumatic balloon actuator is a rubber artificial muscle developed for use in a robot hand in such environments. This actuator has a long stroke, and a high power-to-weight ratio compared with the present pneumatic artificial muscle. Moreover, the dynamic characteristics of this actuator resemble those of human muscle. This study evaluated characteristics of force control of balloon actuator using a predictive functional control (PFC) system with disturbance observer. The predictive functional control is a model-based predictive control (MPC) scheme that predicts the future outputs of the actual plants over the prediction horizon and computes the control effort over the control horizon at every sampling instance. For this study, a 1-link finger system using a pneumatic balloon actuator is developed. Then experiments of PFC control with disturbance observer are performed. These experiments demonstrate the feasibility of its control of a pneumatic balloon actuator for a robot hand.

Keywords: disturbance observer, pneumatic balloon, predictive functional control, rubber artificial muscle

Procedia PDF Downloads 453
2406 Optimizing Electric Vehicle Charging with Charging Data Analytics

Authors: Tayyibah Khanam, Mohammad Saad Alam, Sanchari Deb, Yasser Rafat

Abstract:

Electric vehicles are considered as viable replacements to gasoline cars since they help in reducing harmful emissions and stimulate power generation through renewable energy sources, hence contributing to sustainability. However, one of the significant obstacles in the mass deployment of electric vehicles is the charging time anxiety among users and, thus, the subsequent large waiting times for available chargers at charging stations. Data analytics, on the other hand, has revolutionized the decision-making tasks of management and operating systems since its arrival. In this paper, we attempt to optimize the choice of EV charging stations for users in their vicinity by minimizing the time taken to reach the charging stations and the waiting times for available chargers. Time taken to travel to the charging station is calculated by the Google Maps API and the waiting times are predicted by polynomial regression of the historical data stored. The proposed framework utilizes real-time data and historical data from all operating charging stations in the city and assists the user in finding the best suitable charging station for their current situation and can be implemented in a mobile phone application. The algorithm successfully predicts the most optimal choice of a charging station and the minimum required time for various sample data sets.

Keywords: charging data, electric vehicles, machine learning, waiting times

Procedia PDF Downloads 194
2405 Using T-Splines to Model Point Clouds from Terrestrial Laser Scanner

Authors: G. Kermarrec, J. Hartmann

Abstract:

Spline surfaces are a major representation of freeform surfaces in the computer-aided graphic industry and were recently introduced in the field of geodesy for processing point clouds from terrestrial laser scanner (TLS). The surface fitting consists of approximating a trustworthy mathematical surface to a large numbered 3D point cloud. The standard B-spline surfaces lack of local refinement due to the tensor-product construction. The consequences are oscillating geometry, particularly in the transition from low-to-high curvature parts for scattered point clouds with missing data. More economic alternatives in terms of parameters on how to handle point clouds with a huge amount of observations are the recently introduced T-splines. As long as the partition of unity is guaranteed, their computational complexity is low, and they are flexible. T-splines are implemented in a commercial package called Rhino, a 3D modeler which is widely used in computer aided design to create and animate NURBS objects. We have applied T-splines surface fitting to terrestrial laser scanner point clouds from a bridge under load and a sheet pile wall with noisy observations. We will highlight their potential for modelling details with high trustworthiness, paving the way for further applications in terms of deformation analysis.

Keywords: deformation analysis, surface modelling, terrestrial laser scanner, T-splines

Procedia PDF Downloads 140
2404 Corporate Governance and Business Ethical Values in Organisation: A Study of Unilag Holdings

Authors: Ogunmayi Bamidele, Aribisala Oluwadamilare Olufolarin

Abstract:

The objective of this research was to examine how corporate governance and ethical business values impact both the performance of the organization and its employees, as it is essential for any organization to uphold good ethics and corporate governance. The study was conducted at Unilag Holdings Limited (UniHOLDs) to demonstrate that organizations may experience losses if they do not have proper corporate governance and business ethical values in place. The employees' perception of corporate governance and ethics is crucial for the organization. The research indicated a connection between corporate governance and business ethics values, and therefore, correlation analysis was utilized, making it statistically reliable. The results of the test show a strong positive correlation (r=.812, N=94, P<.01) between corporate governance and business ethical values. A questionnaire was distributed to employees at Unilag Holdings Limited (UniHOLDs), with 94 out of 130 completed and returned. The findings indicate that ethical values contribute to employee productivity, and productive employees have a beneficial impact on the organization's performance. Additionally, the study revealed that employees tend to adhere to rules regardless of their ethical nature. To address this, the organization should ensure that top-level managers do not assign unethical tasks to their subordinates. The study recommends that the organization should consistently practice corporate governance and business ethics. The company needs to make sure that its stakeholders continue to support its way of doing things.

Keywords: business ethical values, corporate governance, organization, business ethics

Procedia PDF Downloads 18
2403 End-to-End Spanish-English Sequence Learning Translation Model

Authors: Vidhu Mitha Goutham, Ruma Mukherjee

Abstract:

The low availability of well-trained, unlimited, dynamic-access models for specific languages makes it hard for corporate users to adopt quick translation techniques and incorporate them into product solutions. As translation tasks increasingly require a dynamic sequence learning curve; stable, cost-free opensource models are scarce. We survey and compare current translation techniques and propose a modified sequence to sequence model repurposed with attention techniques. Sequence learning using an encoder-decoder model is now paving the path for higher precision levels in translation. Using a Convolutional Neural Network (CNN) encoder and a Recurrent Neural Network (RNN) decoder background, we use Fairseq tools to produce an end-to-end bilingually trained Spanish-English machine translation model including source language detection. We acquire competitive results using a duo-lingo-corpus trained model to provide for prospective, ready-made plug-in use for compound sentences and document translations. Our model serves a decent system for large, organizational data translation needs. While acknowledging its shortcomings and future scope, it also identifies itself as a well-optimized deep neural network model and solution.

Keywords: attention, encoder-decoder, Fairseq, Seq2Seq, Spanish, translation

Procedia PDF Downloads 175
2402 The Artificial Intelligence (AI) Impact on Project Management: A Destructive or Transformative Agent

Authors: Kwame Amoah

Abstract:

Artificial intelligence (AI) has the prospect of transforming project management, significantly improving efficiency and accuracy. By automating specific tasks with defined guidelines, AI can assist project managers in making better decisions and allocating resources efficiently, with possible risk mitigation. This study explores how AI is already impacting project management and likely future AI's impact on the field. The AI's reaction has been a divided opinion; while others picture it as a destroyer of jobs, some welcome it as an innovation advocate. Both sides agree that AI will be disruptive and revolutionize PM's functions. If current research is to go by, AI or some form will replace one-third of all learning graduate PM jobs by as early as 2030. A recent survey indicates AI spending will reach $97.9 billion by the end of 2023. Considering such a profound impact, the project management profession will also see a paradigm shift driven by AI. The study examines what the project management profession will look like in the next 5-10 years after this technological disruption. The research methods incorporate existing literature, develop trend analysis, and conduct structured interviews with project management stakeholders from North America to gauge the trend. PM professionals can harness the power of AI, ensuring a smooth transition and positive outcomes. AI adoption will maximize benefits, minimize adverse consequences, and uphold ethical standards, leading to improved project performance.

Keywords: project management, disruptive teacnologies, project management function, AL applications, artificial intelligence

Procedia PDF Downloads 83
2401 Contrastive Learning for Unsupervised Object Segmentation in Sequential Images

Authors: Tian Zhang

Abstract:

Unsupervised object segmentation aims at segmenting objects in sequential images and obtaining the mask of each object without any manual intervention. Unsupervised segmentation remains a challenging task due to the lack of prior knowledge about these objects. Previous methods often require manually specifying the action of each object, which is often difficult to obtain. Instead, this paper does not need action information of objects and automatically learns the actions and relations among objects from the structured environment. To obtain the object segmentation of sequential images, the relationships between objects and images are extracted to infer the action and interaction of objects based on the multi-head attention mechanism. Three types of objects’ relationships in the object segmentation task are proposed: the relationship between objects in the same frame, the relationship between objects in two frames, and the relationship between objects and historical information. Based on these relationships, the proposed model (1) is effective in multiple objects segmentation tasks, (2) just needs images as input, and (3) produces better segmentation results as more relationships are considered. The experimental results on multiple datasets show that this paper’s method achieves state-of-art performance. The quantitative and qualitative analyses of the result are conducted. The proposed method could be easily extended to other similar applications.

Keywords: unsupervised object segmentation, attention mechanism, contrastive learning, structured environment

Procedia PDF Downloads 109
2400 Intrusion Detection in Computer Networks Using a Hybrid Model of Firefly and Differential Evolution Algorithms

Authors: Mohammad Besharatloo

Abstract:

Intrusion detection is an important research topic in network security because of increasing growth in the use of computer network services. Intrusion detection is done with the aim of detecting the unauthorized use or abuse in the networks and systems by the intruders. Therefore, the intrusion detection system is an efficient tool to control the user's access through some predefined regulations. Since, the data used in intrusion detection system has high dimension, a proper representation is required to show the basis structure of this data. Therefore, it is necessary to eliminate the redundant features to create the best representation subset. In the proposed method, a hybrid model of differential evolution and firefly algorithms was employed to choose the best subset of properties. In addition, decision tree and support vector machine (SVM) are adopted to determine the quality of the selected properties. In the first, the sorted population is divided into two sub-populations. These optimization algorithms were implemented on these sub-populations, respectively. Then, these sub-populations are merged to create next repetition population. The performance evaluation of the proposed method is done based on KDD Cup99. The simulation results show that the proposed method has better performance than the other methods in this context.

Keywords: intrusion detection system, differential evolution, firefly algorithm, support vector machine, decision tree

Procedia PDF Downloads 91
2399 Allostatic Load as a Predictor of Adolescents’ Executive Function: A Longitudinal Network Analysis

Authors: Sipu Guo, Silin Huang

Abstract:

Background: Most studies investigate the link between executive function and allostatic load (AL) among adults aged 18 years and older. Studies differed regarding the specific biological indicators studied and executive functions accounted for. Specific executive functions may be differentially related to allostatic load. We investigated the comorbidities of executive functions and allostatic load via network analysis. Methods: We included 603 adolescents (49.84% girls; Mean age = 12.38, SD age = 1.79) from junior high school in rural China. Eight biological markers at T1 and four executive function tasks at T2 were used to evaluate networks. Network analysis was used to determine the network structure, core symptoms, and bridge symptoms in the AL-executive function network among rural adolescents. Results: The executive functions were related to 6 AL biological markers, not to cortisol and epinephrine. The most influential symptoms were inhibition control, cognitive flexibility, processing speed, and systolic blood pressure (SBP). SBP, dehydroepiandrosterone, and processing speed were the bridges through which AL was related to executive functions. dehydroepiandrosterone strongly predicted processing speed. The SBP was the biggest influencer in the entire network. Conclusions: We found evidence for differential relations between markers and executive functions. SBP was a driver in the network; dehydroepiandrosterone showed strong relations with executive function.

Keywords: allostatic load, executive function, network analysis, rural adolescent

Procedia PDF Downloads 52
2398 The Evolution of Amazon Alexa: From Voice Assistant to Smart Home Hub

Authors: Abrar Abuzaid, Maha Alaaeddine, Haya Alesayi

Abstract:

This project is centered around understanding the usage and impact of Alexa, Amazon's popular virtual assistant, in everyday life. Alexa, known for its integration into devices like Amazon Echo, offers functionalities such as voice interaction, media control, providing real-time information, and managing smart home devices. Our primary focus is to conduct a straightforward survey aimed at uncovering how people use Alexa in their daily routines. We plan to reach out to a wide range of individuals to get a diverse perspective on how Alexa is being utilized for various tasks, the frequency and context of its use, and the overall user experience. The survey will explore the most common uses of Alexa, its impact on daily life, features that users find most beneficial, and improvements they are looking for. This project is not just about collecting data but also about understanding the real-world applications of a technology like Alexa and how it fits into different lifestyles. By examining the responses, we aim to gain a practical understanding of Alexa's role in homes and possibly in workplaces. This project will provide insights into user satisfaction and areas where Alexa could be enhanced to meet the evolving needs of its users. It’s a step towards connecting technology with everyday life, making it more accessible and user-friendly

Keywords: Amazon Alexa, artificial intelligence, smart speaker, natural language processing

Procedia PDF Downloads 62
2397 Different Approaches to Teaching a Database Course to Undergraduate and Graduate Students

Authors: Samah Senbel

Abstract:

Database Design is a fundamental part of the Computer Science and Information technology curricula in any school, as well as in the study of management, business administration, and data analytics. In this study, we compare the performance of two groups of students studying the same database design and implementation course at Sacred Heart University in the fall of 2018. Both courses used the same textbook and were taught by the same professor, one for seven graduate students and one for 26 undergraduate students (juniors). The undergraduate students were aged around 20 years old with little work experience, while the graduate students averaged 35 years old and all were employed in computer-related or management-related jobs. The textbook used was 'Database Systems, Design, Implementation, and Management' by Coronel and Morris, and the course was designed to follow the textbook roughly a chapter per week. The first 6 weeks covered the design aspect of a database, followed by a paper exam. The next 6 weeks covered the implementation aspect of the database using SQL followed by a lab exam. Since the undergraduate students are on a 16 week semester, we spend the last three weeks of the course covering NoSQL. This part of the course was not included in this study. After the course was over, we analyze the results of the two groups of students. An interesting discrepancy was observed: In the database design part of the course, the average grade of the graduate students was 92%, while that of the undergraduate students was 77% for the same exam. In the implementation part of the course, we observe the opposite: the average grade of the graduate students was 65% while that of the undergraduate students was 73%. The overall grades were quite similar: the graduate average was 78% and that of the undergraduates was 75%. Based on these results, we concluded that having both classes follow the same time schedule was not beneficial, and an adjustment is needed. The graduates could spend less time on design and the undergraduates would benefit from more design time. In the fall of 2019, 30 students registered for the undergraduate course and 15 students registered for the graduate course. To test our conclusion, the undergraduates spend about 67% of time (eight classes) on the design part of the course and 33% (four classes) on the implementation part, using the exact exams as the previous year. This resulted in an improvement in their average grades on the design part from 77% to 83% and also their implementation average grade from 73% to 79%. In conclusion, we recommend using two separate schedules for teaching the database design course. For undergraduate students, it is important to spend more time on the design part rather than the implementation part of the course. While for the older graduate students, we recommend spending more time on the implementation part, as it seems that is the part they struggle with, even though they have a higher understanding of the design component of databases.

Keywords: computer science education, database design, graduate and undergraduate students, pedagogy

Procedia PDF Downloads 121
2396 Empowering Rural Women Entrepreneurs via Microcredit

Authors: Salwana Hassan, Rashidah Abdul Rahman

Abstract:

Poverty in rural Malaysia remains unresolved and contribute 7.8% to the whole poverty figure in Malaysia. Among the rural folks, 50% is women. Thus, women, as the significant human capital to fight the long lost battle of poverty , is indispensable. This will also serve as an equal opportunity for women to play active and positive roles to develop the society that has been the tasks for men all this while. More importantly rural women folks have the potential to offer better quality of life for their family by providing extra income and monetary support whenever their husbands are not able to work. The reality in this, however, cannot be solved easily as there are many factors that stand in the way and prevent the resolutions to be observed.In this regard, this paper describes a model that has been used to resolve such issues in rural Malaysia. The model utilizes a synergetic effort between an academic institution, an NGO that govern the rural women folks and a private trading company that sell the finished product. The project was conducted in rural area of Selangor and has been in operations since the end of 2013. It shows positive outcome and could be used in other rural areas of Malaysia. The project captures the influence of the NGO programs upon rural women entrepreneurship and how a private trading company can facilitate to help develop a community. As a result the project reveals that self-income generating activities by entrepreneurship are the important contributing factor to empowering rural women folks in Malaysia.

Keywords: poverty, empowerment, rural, entrepreneurship, community

Procedia PDF Downloads 393
2395 Analyzing Students' Writing in an English Code-Mixing Context in Nepali: An Ecological and Systematic Functional Approach

Authors: Binod Duwadi

Abstract:

This article examines the language and literacy practices of English Code-mixing in Nepalese Classroom. Situating the study within an ecological framework, a systematic functional linguistic (SFL) approach was used to analyze students writing in two Neplease schools. Data collection included interviews with teachers, classroom observations, instructional materials, and focal students’ writing samples. Data analyses revealed vastly different language ecologies between the schools owing to sharp socioeconomic stratification, the structural organization of schools, and the pervasiveness of standard language ideology, with stigmatizes English code mixing (ECM) and privileges Standard English in schools. Functional analysis of students’ writing showed that the nature of the writing tasks at the schools created different affordances for exploiting lexicogrammatically choices for meaning making-enhancing them in the case of one school but severely restricting them in the case of another- perpetuating the academic disadvantage for code mixing speakers. Recommendations for structural and attitudinal changes through teacher training and implementation of approaches that engage students’ bidialectal competence for learning are made as important first steps towards addressing educational inequities in Nepalese schools.

Keywords: code-mixing, ecological perspective, systematic functional approach, language and identity

Procedia PDF Downloads 124
2394 A Novel Exploration/Exploitation Policy Accelerating Learning In Both Stationary And Non Stationary Environment Navigation Tasks

Authors: Wiem Zemzem, Moncef Tagina

Abstract:

In this work, we are addressing the problem of an autonomous mobile robot navigating in a large, unknown and dynamic environment using reinforcement learning abilities. This problem is principally related to the exploration/exploitation dilemma, especially the need to find a solution letting the robot detect the environmental change and also learn in order to adapt to the new environmental form without ignoring knowledge already acquired. Firstly, a new action selection strategy, called ε-greedy-MPA (the ε-greedy policy favoring the most promising actions) is proposed. Unlike existing exploration/exploitation policies (EEPs) such as ε-greedy and Boltzmann, the new EEP doesn’t only rely on the information of the actual state but also uses those of the eventual next states. Secondly, as the environment is large, an exploration favoring least recently visited states is added to the proposed EEP in order to accelerate learning. Finally, various simulations with ball-catching problem have been conducted to evaluate the ε-greedy-MPA policy. The results of simulated experiments show that combining this policy with the Qlearning method is more effective and efficient compared with the ε-greedy policy in stationary environments and the utility-based reinforcement learning approach in non stationary environments.

Keywords: autonomous mobile robot, exploration/ exploitation policy, large, dynamic environment, reinforcement learning

Procedia PDF Downloads 417