Search results for: large language models (LLMS)
14940 Nanda Ways of Knowing, Being and Doing: Our Process of Research Engagement and Research Impacts
Authors: Steven Kelly
Abstract:
A fundament role of the researcher is research engagement, that is, the interaction between researchers and research end-users outside of academia for the mutually beneficial transfer of knowledge, technologies, methods, or resources. While research impact is the contribution that research makes to the economy, society, environment, or culture beyond the contribution to academic research. Ironically, traditional impact metrics in the academy are designed to focus on the outputs; it dismisses the important role engagement plays in fostering a collaborative process that leads to meaningful, ethical, and useful impacts. Dr. Kelly, aNanda (First Nations) man himself, has worked closely with the Nanda community over the past decade, ensuring cultural protocols are upheld and implemented while doing research engagement. The focus was on the process, which was essential to foster a positive research impact culture. The contributions that flowed from this process were the naming of a new species of squat lobster in the Nanda language, a poster design in collaboration with The University of Melbourne, Museums Victoria and Bundiyarra - IrraWanga language centre, media coverage, and the formation of the “Nanda language, Nanda country project”. The Nanda language, Nanda country project is a language revitalization project that focused on reconnecting Nanda people with the language & culture on Nanda Country. Such outcomes are imperative on the eve of the United Nations International Decade of Indigenous Languages. In this paperDr, Kellywill discuss howNanda cultural practicesinformed research engagement to foster a collaborative processthat, in turn, ledto meaningful, ethical, and useful impacts within and outside of the academy.Keywords: community collaboration, indigenous, nanda, research engagement, research impacts
Procedia PDF Downloads 11414939 Structure Design of Vacuum Vessel with Large Openings for Spacecraft Thermal Vacuum Test
Authors: Han Xiao, Ruan Qi, Zhang Lei, Qi Yan
Abstract:
Space environment simulator is a facility used to conduct thermal test for spacecraft, and vacuum vessel is the main body of it. According to the requirements for thermal tests of the spacecraft and its solar array panels, the primary vessel and the side vessels are designed to be a combinative structure connected with aperture, which ratio reaches 0.7. Since the vacuum vessel suffers 0.1MPa external pressure during the process of thermal test, in order to ensure the simulator’s reliability and safety, it’s necessary to calculate the vacuum vessel’s intensity and stability. Based on the impact of large openings to vacuum vessel structure, this paper explored the reinforce design and analytical way of vacuum vessel with large openings, using a large space environment simulator’s vacuum vessel design as an example. Tests showed that the reinforce structure is effective to fulfill the requirements of external pressure and the gravity. This ensured the reliability of the space environment simulator, providing a guarantee for developing the spacecraft.Keywords: vacuum vessel, large opening, space environment simulator, structure design
Procedia PDF Downloads 53514938 A Comparative Analysis of ARIMA and Threshold Autoregressive Models on Exchange Rate
Authors: Diteboho Xaba, Kolentino Mpeta, Tlotliso Qejoe
Abstract:
This paper assesses the in-sample forecasting of the South African exchange rates comparing a linear ARIMA model and a SETAR model. The study uses a monthly adjusted data of South African exchange rates with 420 observations. Akaike information criterion (AIC) and the Schwarz information criteria (SIC) are used for model selection. Mean absolute error (MAE), root mean squared error (RMSE) and mean absolute percentage error (MAPE) are error metrics used to evaluate forecast capability of the models. The Diebold –Mariano (DM) test is employed in the study to check forecast accuracy in order to distinguish the forecasting performance between the two models (ARIMA and SETAR). The results indicate that both models perform well when modelling and forecasting the exchange rates, but SETAR seemed to outperform ARIMA.Keywords: ARIMA, error metrices, model selection, SETAR
Procedia PDF Downloads 24414937 A Trend Based Forecasting Framework of the ATA Method and Its Performance on the M3-Competition Data
Authors: H. Taylan Selamlar, I. Yavuz, G. Yapar
Abstract:
It is difficult to make predictions especially about the future and making accurate predictions is not always easy. However, better predictions remain the foundation of all science therefore the development of accurate, robust and reliable forecasting methods is very important. Numerous number of forecasting methods have been proposed and studied in the literature. There are still two dominant major forecasting methods: Box-Jenkins ARIMA and Exponential Smoothing (ES), and still new methods are derived or inspired from them. After more than 50 years of widespread use, exponential smoothing is still one of the most practically relevant forecasting methods available due to their simplicity, robustness and accuracy as automatic forecasting procedures especially in the famous M-Competitions. Despite its success and widespread use in many areas, ES models have some shortcomings that negatively affect the accuracy of forecasts. Therefore, a new forecasting method in this study will be proposed to cope with these shortcomings and it will be called ATA method. This new method is obtained from traditional ES models by modifying the smoothing parameters therefore both methods have similar structural forms and ATA can be easily adapted to all of the individual ES models however ATA has many advantages due to its innovative new weighting scheme. In this paper, the focus is on modeling the trend component and handling seasonality patterns by utilizing classical decomposition. Therefore, ATA method is expanded to higher order ES methods for additive, multiplicative, additive damped and multiplicative damped trend components. The proposed models are called ATA trended models and their predictive performances are compared to their counter ES models on the M3 competition data set since it is still the most recent and comprehensive time-series data collection available. It is shown that the models outperform their counters on almost all settings and when a model selection is carried out amongst these trended models ATA outperforms all of the competitors in the M3- competition for both short term and long term forecasting horizons when the models’ forecasting accuracies are compared based on popular error metrics.Keywords: accuracy, exponential smoothing, forecasting, initial value
Procedia PDF Downloads 17714936 A Motion Dictionary to Real-Time Recognition of Sign Language Alphabet Using Dynamic Time Warping and Artificial Neural Network
Authors: Marcio Leal, Marta Villamil
Abstract:
Computacional recognition of sign languages aims to allow a greater social and digital inclusion of deaf people through interpretation of their language by computer. This article presents a model of recognition of two of global parameters from sign languages; hand configurations and hand movements. Hand motion is captured through an infrared technology and its joints are built into a virtual three-dimensional space. A Multilayer Perceptron Neural Network (MLP) was used to classify hand configurations and Dynamic Time Warping (DWT) recognizes hand motion. Beyond of the method of sign recognition, we provide a dataset of hand configurations and motion capture built with help of fluent professionals in sign languages. Despite this technology can be used to translate any sign from any signs dictionary, Brazilian Sign Language (Libras) was used as case study. Finally, the model presented in this paper achieved a recognition rate of 80.4%.Keywords: artificial neural network, computer vision, dynamic time warping, infrared, sign language recognition
Procedia PDF Downloads 21714935 Taleghan Dam Break Numerical Modeling
Authors: Hamid Goharnejad, Milad Sadeghpoor Moalem, Mahmood Zakeri Niri, Leili Sadeghi Khalegh Abadi
Abstract:
While there are many benefits to using reservoir dams, their break leads to destructive effects. From the viewpoint of International Committee of Large Dams (ICOLD), dam break means the collapse of whole or some parts of a dam; thereby the dam will be unable to hold water. Therefore, studying dam break phenomenon and prediction of its behavior and effects reduces losses and damages of the mentioned phenomenon. One of the most common types of reservoir dams is embankment dam. Overtopping in embankment dams occurs because of flood discharge system inability in release inflows to reservoir. One of the most important issues among managers and engineers to evaluate the performance of the reservoir dam rim when sliding into the storage, creating waves is large and long. In this study, the effects of floods which caused the overtopping of the dam have been investigated. It was assumed that spillway is unable to release the inflow. To determine outflow hydrograph resulting from dam break, numerical model using Flow-3D software and empirical equations was used. Results of numerical models and their comparison with empirical equations show that numerical model and empirical equations can be used to study the flood resulting from dam break.Keywords: embankment dam break, empirical equations, Taleghan dam, Flow-3D numerical model
Procedia PDF Downloads 32114934 Refugees’inclusion: The Psychological Screening and the Educational Tools in Portugal
Authors: Sandra Figueiredo
Abstract:
To guarantee the well-being and the academic achievement it is crucial into the global society to develop techniques to assess language competence and control psychological aspects on the second language learning context. The current scenario of the war conflicts that are emerging mostly in Europe and Middle East have been resulting in forced immigration and refugees’ maladjustment. The inclusion is the priority for United Nations concerning the sustainability of societies. For inclusion, psychological screening tests and educational tools are urgent. Method: Approximately 100 refugees from Ukraine were assessed, in Portugal, under the administration of the PCL-5. This 20-item instrument evaluates the Post-Traumatic Disorder. Expected results: The statistical analysis will be performed with the International Database Analyzer and SPSS (v. 28). The results expected are the relationship between traumatic events caused by war and post-traumatic symptomatology (anxiety, hypervigilance, stress). Implications: The data will be discussed concerning the problems of belonging, the psychological constraints and educational attainment (language needs included) experienced by the individuals more recently arrived to the hosting societies. The refugees’ acculturation process and the emotional regulation will be addressed.Keywords: refugees, immigration, educational needs, trauma, inclusion, second language.
Procedia PDF Downloads 5814933 A Computational Analysis of Flow and Acoustics around a Car Wing Mirror
Authors: Aidan J. Bowes, Reaz Hasan
Abstract:
The automotive industry is continually aiming to develop the aerodynamics of car body design. This may be for a variety of beneficial reasons such as to increase speed or fuel efficiency by reducing drag. However recently there has been a greater amount of focus on wind noise produced while driving. Designers in this industry seek a combination of both simplicity of approach and overall effectiveness. This combined with the growing availability of commercial CFD (Computational Fluid Dynamics) packages is likely to lead to an increase in the use of RANS (Reynolds Averaged Navier-Stokes) based CFD methods. This is due to these methods often being simpler than other CFD methods, having a lower demand on time and computing power. In this investigation the effectiveness of turbulent flow and acoustic noise prediction using RANS based methods has been assessed for different wing mirror geometries. Three different RANS based models were used, standard k-ε, realizable k-ε and k-ω SST. The merits and limitations of these methods are then discussed, by comparing with both experimental and numerical results found in literature. In general, flow prediction is fairly comparable to more complex LES (Large Eddy Simulation) based methods; in particular for the k-ω SST model. However acoustic noise prediction still leaves opportunities for more improvement using RANS based methods.Keywords: acoustics, aerodynamics, RANS models, turbulent flow
Procedia PDF Downloads 44614932 Reflections on Opportunities and Challenges for Systems Engineering
Authors: Ali E. Abbas
Abstract:
This paper summarizes some of the discussions that occurred in a workshop in West Virginia, U.S.A which was sponsored by the National Science Foundation (NSF) in February 2016. The goal of the workshop was to explore the opportunities and challenges for applying systems engineering in large enterprises, and some of the issues that still persist. The main topics of the discussion included challenges with elaboration and abstraction in large systems, interfacing physical and social systems, and the need for axiomatic frameworks for large enterprises. We summarize these main points of discussion drawing parallels with decision making in organizations to instigate research in these discussion areas.Keywords: decision analysis, systems engineering, framing, value creation
Procedia PDF Downloads 32914931 Advancing Communication Theory in the Age of Digital Technology: Bridging the Gap Between Traditional Models and Emerging Platforms
Authors: Sidique Fofanah
Abstract:
This paper explores the intersection of traditional communication theories and modern digital technologies, analyzing how established models adapt to contemporary communication platforms. It examines the evolving nature of interpersonal, group, and mass communication within digital environments, emphasizing the role of social media, AI-driven communication tools, and virtual reality in reshaping communication paradigms. The paper also discusses the implications for future research and practice in communication studies, proposing an integrated framework that accommodates both classical and emerging theories.Keywords: communication, traditional models, emerging platforms, digital media
Procedia PDF Downloads 2614930 Mathematical Modeling of Carotenoids and Polyphenols Content of Faba Beans (Vicia faba L.) during Microwave Treatments
Authors: Ridha Fethi Mechlouch, Ahlem Ayadi, Ammar Ben Brahim
Abstract:
Given the importance of the preservation of polyphenols and carotenoids during thermal processing, we attempted in this study to investigate the variation of these two parameters in faba beans during microwave treatment using different power densities (1; 2; and 3W/g), then to perform a mathematical modeling by using non-linear regression analysis to evaluate the models constants. The variation of the carotenoids and polyphenols ratio of faba beans and the models are tested to validate the experimental results. Exponential models were found to be suitable to describe the variation of caratenoid ratio (R²= 0.945, 0.927 and 0.946) for power densities (1; 2; and 3W/g) respectively, and polyphenol ratio (R²= 0.931, 0.989 and 0.982) for power densities (1; 2; and 3W/g) respectively. The effect of microwave power density Pd(W/g) on the coefficient k of models were also investigated. The coefficient is highly correlated (R² = 1) and can be expressed as a polynomial function.Keywords: microwave treatment, power density, carotenoid, polyphenol, modeling
Procedia PDF Downloads 25914929 The Sapir-Whorf Hypothesis and Multicultural Effects on Translators: A Case Study from Chinese Ethnic Minority Literature
Authors: Yuqiao Zhou
Abstract:
The Sapir-Whorf hypothesis (SWH) emphasizes the effect produced by language on people’s minds. According to linguistic relativity, language has evolved over the course of human life on earth, and, in turn, the acquisition of language shapes learners’ thoughts. Despite much attention drawn by SWH, few scholars have attempted to analyse people’s thoughts via their literary works. And yet, the linguistic choices that create a narrative can enable us to examine its writer’s thoughts. Still, less work has been done on the impact of language on the minds of bilingual people. Internationalization has resulted in an increasing number of bilingual and multilingual individuals. In China, where more than one hundred languages are used for communication, most people are bilingual in Mandarin Chinese (the official language of China) and their own dialect. Taking as its corpus the ethnic minority myth of Ge Sa-er Wang by Alai and its English translation by Goldblatt and Lin, this paper aims to analyse the effects of culture on bilingual people’s minds. It will first analyse Alai’s thoughts on using the original version of Ge Sa-er Wang; next, it will examine the thoughts of the two translators by looking at translation choices made in the English version; finally, it will compare the cultural influences evident in the thoughts of Alai, and Goldblatt and Lin. Whereas Alai can speak two Sino-Tibetan languages – Mandarin Chinese and Tibetan – Goldblatt and Lin can speak two languages from different families – Mandarin Chinese (a Sino-Tibetan language) and English (an Indo-European language). The results reveal two systems of thought existing in the translators’ minds; Alai’s text, on the other hand, does not reveal a significant influence from North China, where Mandarin Chinese originated. The findings reveal the inconsistency of a second language’s influence on people’s minds. Notably, they suggest that the more different the two languages are, the greater the influence produced by the second language culture on people’s thoughts. It is hoped that this research will expand the scope of SWH as well as shed light on future translation studies on ethnic minority literature.Keywords: Sapir-Whorf hypothesis, cultural translation, cultural-specific items, Ge Sa-er Wang, ethnic minority literature, Tibet
Procedia PDF Downloads 11514928 Exchange Rate Forecasting by Econometric Models
Authors: Zahid Ahmad, Nosheen Imran, Nauman Ali, Farah Amir
Abstract:
The objective of the study is to forecast the US Dollar and Pak Rupee exchange rate by using time series models. For this purpose, daily exchange rates of US and Pakistan for the period of January 01, 2007 - June 2, 2017, are employed. The data set is divided into in sample and out of sample data set where in-sample data are used to estimate as well as forecast the models, whereas out-of-sample data set is exercised to forecast the exchange rate. The ADF test and PP test are used to make the time series stationary. To forecast the exchange rate ARIMA model and GARCH model are applied. Among the different Autoregressive Integrated Moving Average (ARIMA) models best model is selected on the basis of selection criteria. Due to the volatility clustering and ARCH effect the GARCH (1, 1) is also applied. Results of analysis showed that ARIMA (0, 1, 1 ) and GARCH (1, 1) are the most suitable models to forecast the future exchange rate. Further the GARCH (1,1) model provided the volatility with non-constant conditional variance in the exchange rate with good forecasting performance. This study is very useful for researchers, policymakers, and businesses for making decisions through accurate and timely forecasting of the exchange rate and helps them in devising their policies.Keywords: exchange rate, ARIMA, GARCH, PAK/USD
Procedia PDF Downloads 56114927 Study on Flexible Diaphragm In-Plane Model of Irregular Multi-Storey Industrial Plant
Authors: Cheng-Hao Jiang, Mu-Xuan Tao
Abstract:
The rigid diaphragm model may cause errors in the calculation of internal forces due to neglecting the in-plane deformation of the diaphragm. This paper thus studies the effects of different diaphragm in-plane models (including in-plane rigid model and in-plane flexible model) on the seismic performance of structures. Taking an actual industrial plant as an example, the seismic performance of the structure is predicted using different floor diaphragm models, and the analysis errors caused by different diaphragm in-plane models including deformation error and internal force error are calculated. Furthermore, the influence of the aspect ratio on the analysis errors is investigated. Finally, the code rationality is evaluated by assessing the analysis errors of the structure models whose floors were determined as rigid according to the code’s criterion. It is found that different floor models may cause great differences in the distribution of structural internal forces, and the current code may underestimate the influence of the floor in-plane effect.Keywords: industrial plant, diaphragm, calculating error, code rationality
Procedia PDF Downloads 14014926 Recommender System Based on Mining Graph Databases for Data-Intensive Applications
Authors: Mostafa Gamal, Hoda K. Mohamed, Islam El-Maddah, Ali Hamdi
Abstract:
In recent years, many digital documents on the web have been created due to the rapid growth of ’social applications’ communities or ’Data-intensive applications’. The evolution of online-based multimedia data poses new challenges in storing and querying large amounts of data for online recommender systems. Graph data models have been shown to be more efficient than relational data models for processing complex data. This paper will explain the key differences between graph and relational databases, their strengths and weaknesses, and why using graph databases is the best technology for building a realtime recommendation system. Also, The paper will discuss several similarity metrics algorithms that can be used to compute a similarity score of pairs of nodes based on their neighbourhoods or their properties. Finally, the paper will discover how NLP strategies offer the premise to improve the accuracy and coverage of realtime recommendations by extracting the information from the stored unstructured knowledge, which makes up the bulk of the world’s data to enrich the graph database with this information. As the size and number of data items are increasing rapidly, the proposed system should meet current and future needs.Keywords: graph databases, NLP, recommendation systems, similarity metrics
Procedia PDF Downloads 10414925 Evaluate the Possibility of Using ArcGIS Basemaps as GCP for Large Scale Maps
Authors: Jali Octariady, Ida Herliningsih, Ade K. Mulyana, Annisa Fitria, Diaz C. K. Yuwana
Abstract:
Awareness of the importance large-scale maps for development of a country is growing in all walks of life, especially for governments in Indonesia. Various parties, especially local governments throughout Indonesia demanded for immediate availability the large-scale maps of 1:5000 for regional development. But in fact, the large-scale maps of 1:5000 is only available less than 5% of the entire territory of Indonesia. Unavailability precise GCP at the entire territory of Indonesia is one of causes of slow availability the large scale maps of 1:5000. This research was conducted to find an alternative solution to this problem. This study was conducted to assess the accuracy of ArcGIS base maps coordinate when it shall be used as GCP for creating a map scale of 1:5000. The study was conducted by comparing the GCP coordinate from Field survey using GPS Geodetic than the coordinate from ArcGIS basemaps in various locations in Indonesia. Some areas are used as a study area are Lombok Island, Kupang City, Surabaya City and Kediri District. The differences value of the coordinates serve as the basis for assessing the accuracy of ArcGIS basemaps coordinates. The results of the study at various study area show the variation of the amount of the coordinates value given. Differences coordinate value in the range of millimeters (mm) to meters (m) in the entire study area. This is shown the inconsistency quality of ArcGIS base maps coordinates. This inconsistency shows that the coordinate value from ArcGIS Basemaps is careless. The Careless coordinate from ArcGIS Basemaps indicates that its cannot be used as GCP for large-scale mapping on the entire territory of Indonesia.Keywords: accuracy, ArcGIS base maps, GCP, large scale maps
Procedia PDF Downloads 37314924 Multiple Identity Construction among Multilingual Minorities: A Quantitative Sociolinguistic Case Study
Authors: Stefanie Siebenhütter
Abstract:
This paper aims to reveal criterions involved in the process of identity-forming among multilingual minority language speakers in Northeastern Thailand and in the capital Bangkok. Using sociolinguistic interviews and questionnaires, it is asked which factors are important for speakers and how they define their identity by their interactions socially as well as linguistically. One key question to answer is how sociolinguistic factors may force or diminish the process of forming social identity of multilingual minority speakers. However, the motivation for specific language use is rarely overt to the speaker’s themselves as well as to others. Therefore, identifying the intentions included in the process of identity construction is to approach by scrutinizing speaker’s behavior and attitudes. Combining methods used in sociolinguistics and social psychology allows uncovering the tools for identity construction that ethnic Kui uses to range themselves within a multilingual setting. By giving an overview of minority speaker’s language use in context of the specific border near multilingual situation and asking how speakers construe identity within this spatial context, the results exhibit some of the subtle and mostly unconscious criterions involved in the ongoing process of identity construction.Keywords: social identity, identity construction, minority language, multilingualism, social networks, social boundaries
Procedia PDF Downloads 26714923 Virtual Learning during the Period of COVID-19 Pandemic at a Saudi University
Authors: Ahmed Mohammed Omer Alghamdi
Abstract:
Since the COVID-19 pandemic started, a rapid, unexpected transition from face-to-face to virtual classroom (VC) teaching has involved several challenges and obstacles. However, there are also opportunities and thoughts that need to be examined and discussed. In addition, the entire world is witnessing that the teaching system and, more particularly, higher education institutes have been interrupted. To maintain the learning and teaching practices as usual, countries were forced to transition from traditional to virtual classes using various technology-based devices. In this regard, the Kingdom of Saudi Arabia (KSA) is no exception. Focusing on how the current situation has forced many higher education institutes to change to virtual classes may possibly provide a clear insight into adopted practices and implications. The main purpose of this study, therefore, was to investigate how both Saudi English as a foreign language (EFL) teachers and students perceived the implementation of virtual classes as a key factor for useful language teaching and learning process during the COVID-19 pandemic period at a Saudi university. The impetus for the research was, therefore, the need to find ways of identifying the deficiencies in this application and to suggest possible solutions that might rectify those deficiencies. This study seeks to answer the following overarching research question: “How do Saudi EFL instructors and students perceive the use of virtual classes during the COVID-19 pandemic period in their language teaching and learning context?” The following sub-questions are also used to guide the design of the study to answer the main research question: (1) To what extent are virtual classes important intra-pandemic from Saudi EFL instructors’ and students’ perspectives? (2) How effective are virtual classes for fostering English language students’ achievement? (3) What are the challenges and obstacles that instructors and students may face during the implementation of virtual teaching? A mixed method approach was employed in this study; the questionnaire data collection represented the quantitative method approach for this study, whereas the transcripts of recorded interviews represented the qualitative method approach. The participants included EFL teachers (N = 4) and male and female EFL students (N = 36). Based on the findings of this study, various aspects from teachers' and students’ perspectives were examined to determine the use of the virtual classroom applications in terms of fulfilling the students’ English language learning needs. The major findings of the study revealed that the virtual classroom applications during the current pandemic situation encountered three major challenges, among which the existence of the following essential aspects, namely lack of technology and an internet connection, having a large number of students in a virtual classroom and lack of students’ and teachers’ interactions during the virtual classroom applications. Finally, the findings indicated that although Saudi EFL students and teachers view the virtual classrooms in a positive light during the pandemic period, they reported that for long and post-pandemic period, they preferred the traditional face-to-face teaching procedure.Keywords: virtual classes, English as a foreign language, COVID-19, Internet, pandemic
Procedia PDF Downloads 8614922 Application Difference between Cox and Logistic Regression Models
Authors: Idrissa Kayijuka
Abstract:
The logistic regression and Cox regression models (proportional hazard model) at present are being employed in the analysis of prospective epidemiologic research looking into risk factors in their application on chronic diseases. However, a theoretical relationship between the two models has been studied. By definition, Cox regression model also called Cox proportional hazard model is a procedure that is used in modeling data regarding time leading up to an event where censored cases exist. Whereas the Logistic regression model is mostly applicable in cases where the independent variables consist of numerical as well as nominal values while the resultant variable is binary (dichotomous). Arguments and findings of many researchers focused on the overview of Cox and Logistic regression models and their different applications in different areas. In this work, the analysis is done on secondary data whose source is SPSS exercise data on BREAST CANCER with a sample size of 1121 women where the main objective is to show the application difference between Cox regression model and logistic regression model based on factors that cause women to die due to breast cancer. Thus we did some analysis manually i.e. on lymph nodes status, and SPSS software helped to analyze the mentioned data. This study found out that there is an application difference between Cox and Logistic regression models which is Cox regression model is used if one wishes to analyze data which also include the follow-up time whereas Logistic regression model analyzes data without follow-up-time. Also, they have measurements of association which is different: hazard ratio and odds ratio for Cox and logistic regression models respectively. A similarity between the two models is that they are both applicable in the prediction of the upshot of a categorical variable i.e. a variable that can accommodate only a restricted number of categories. In conclusion, Cox regression model differs from logistic regression by assessing a rate instead of proportion. The two models can be applied in many other researches since they are suitable methods for analyzing data but the more recommended is the Cox, regression model.Keywords: logistic regression model, Cox regression model, survival analysis, hazard ratio
Procedia PDF Downloads 45514921 Under the Veneer of Words Lies Power: Foucauldian Analysis of Oleanna
Authors: Diba Arjmandi
Abstract:
The notion of power and gender domination is one of the inseparable aspects of themes in postmodern literature. The reason of its importance has been discussed frequently since the rise of Michel Foucault and his vantage point toward the circulation of power and the transgression of forces. The language and society act as the basic grounds for the study, as all human beings are bound to the set of rules and norms which shape them in the acceptable way in the macrocosm. How different genders in different positions behave and show reactions to the provocation of social forces and superiority of one another, is of great interest to writers and literary critics. Mamet’s works are noticeable for their controversial but timely themes which illustrate the human conflict with the community and greed for power. Many critics like Christopher Bigsby and Harold Bloom have been discussing Mamet and his ideas during recent years. This paper is the study of Oleanna, Mamet’s masterpiece about teacher-student relationship and the circulation of power between a man and woman. He shows the very breakable boundaries in domination of a gender and the downfall of speech as the consequence of transgression and freedom. The failure of the language the teacher uses and the abuses of his own words by a student who seeks superiority and knowledge are the main subjects of discussion. Supported by the ideas of Foucault, the language Mamet uses to represent his characters becomes the fundamental element of this survey. As a result, language becomes both the means of achievement and also downfall.Keywords: domination, foucault, language, mamet, oleanna, power, transgression
Procedia PDF Downloads 48514920 Testing the Simplification Hypothesis in Constrained Language Use: An Entropy-Based Approach
Authors: Jiaxin Chen
Abstract:
Translations have been labeled as more simplified than non-translations, featuring less diversified and more frequent lexical items and simpler syntactic structures. Such simplified linguistic features have been identified in other bilingualism-influenced language varieties, including non-native and learner language use. Therefore, it has been proposed that translation could be studied within a broader framework of constrained language, and simplification is one of the universal features shared by constrained language varieties due to similar cognitive-physiological and social-interactive constraints. Yet contradicting findings have also been presented. To address this issue, this study intends to adopt Shannon’s entropy-based measures to quantify complexity in language use. Entropy measures the level of uncertainty or unpredictability in message content, and it has been adapted in linguistic studies to quantify linguistic variance, including morphological diversity and lexical richness. In this study, the complexity of lexical and syntactic choices will be captured by word-form entropy and pos-form entropy, and a comparison will be made between constrained and non-constrained language use to test the simplification hypothesis. The entropy-based method is employed because it captures both the frequency of linguistic choices and their evenness of distribution, which are unavailable when using traditional indices. Another advantage of the entropy-based measure is that it is reasonably stable across languages and thus allows for a reliable comparison among studies on different language pairs. In terms of the data for the present study, one established (CLOB) and two self-compiled corpora will be used to represent native written English and two constrained varieties (L2 written English and translated English), respectively. Each corpus consists of around 200,000 tokens. Genre (press) and text length (around 2,000 words per text) are comparable across corpora. More specifically, word-form entropy and pos-form entropy will be calculated as indicators of lexical and syntactical complexity, and ANOVA tests will be conducted to explore if there is any corpora effect. It is hypothesized that both L2 written English and translated English have lower entropy compared to non-constrained written English. The similarities and divergences between the two constrained varieties may provide indications of the constraints shared by and peculiar to each variety.Keywords: constrained language use, entropy-based measures, lexical simplification, syntactical simplification
Procedia PDF Downloads 9414919 Suburban Large Residential Area Development Strategy with an Example of Liangzhu Culture Village in Hangzhou
Authors: Liang Fang
Abstract:
The development of the large suburban residential area is a product of the leap development during the rapid urbanization process in China. On the process of the large-scale development of large settlements in a short time, various problems arose in the suburban residential area, such as spatial layout being disorder, basic facilities construction lagging behind and being unreasonable, residential neighborhood space and street culture missing. Aimed at the contradictions mentioned above, exploring a way is imminent to construct appropriate residential area. We select a typical Liangzhu Culture Village in Hangzhou and put forward functional composite residential area of fine development strategy, along which business promotes and assists community autonomy and then a good community culture is constructed. All in all, the development and construction mode, contributing to an all-people and full-time participation, is beneficial to create a harmonious community of sustainable development, which gives good implication to a single enterprise development city real estate projects.Keywords: community autonomy, development and construction mode, functional composite, suburban large residential area
Procedia PDF Downloads 35814918 From User's Requirements to UML Class Diagram
Authors: Zeineb Ben Azzouz, Wahiba Ben Abdessalem Karaa
Abstract:
The automated extraction of UML class diagram from natural language requirements is a highly challenging task. Many approaches, frameworks and tools have been presented in this field. Nonetheless, the experiments of these tools have shown that there is no approach that can work best all the time. In this context, we propose a new accurate approach to facilitate the automatic mapping from textual requirements to UML class diagram. Our new approach integrates the best properties of statistical Natural Language Processing (NLP) techniques to reduce ambiguity when analysing natural language requirements text. In addition, our approach follows the best practices defined by conceptual modelling experts to determine some patterns indispensable for the extraction of basic elements and concepts of the class diagram. Once the relevant information of class diagram is captured, a XMI document is generated and imported with a CASE tool to build the corresponding UML class diagram.Keywords: class diagram, user’s requirements, XMI, software engineering
Procedia PDF Downloads 47114917 Analysis of Moving Loads on Bridges Using Surrogate Models
Authors: Susmita Panda, Arnab Banerjee, Ajinkya Baxy, Bappaditya Manna
Abstract:
The design of short to medium-span high-speed bridges in critical locations is an essential aspect of vehicle-bridge interaction. Due to dynamic interaction between moving load and bridge, mathematical models or finite element modeling computations become time-consuming. Thus, to reduce the computational effort, a universal approximator using an artificial neural network (ANN) has been used to evaluate the dynamic response of the bridge. The data set generation and training of surrogate models have been conducted over the results obtained from mathematical modeling. Further, the robustness of the surrogate model has been investigated, which showed an error percentage of less than 10% with conventional methods. Additionally, the dependency of the dynamic response of the bridge on various load and bridge parameters has been highlighted through a parametric study.Keywords: artificial neural network, mode superposition method, moving load analysis, surrogate models
Procedia PDF Downloads 10014916 Variation in Complement Order in English: Implications for Interlanguage Syntax
Authors: Juliet Udoudom
Abstract:
Complement ordering principles of natural language phrases (XPs) stipulate that Head terms be consistently placed phrase initially or phrase-finally, yielding two basic theoretical orders – Head – Complement order or Complement – Head order. This paper examines the principles which determine complement ordering in English V- and N-bar structures. The aim is to determine the extent to which complement linearisations in the two phrase types are consistent with the two theoretical orders outlined above given the flexible and varied nature of natural language structures. The objective is to see whether there are variation(s) in the complement linearisations of the XPs studied and the implications which such variations hold for the inter-language syntax of English and Ibibio. A corpus-based approach was employed in obtaining the English data. V- and -N – bar structures containing complement structures were isolated for analysis. Data were examined from the perspective of the X-bar and Government – theories of Chomsky’s (1981) Government-Binding format. Findings from the analysis show that in V – bar structures in English, heads are consistently placed phrase – initially yielding a Head – Complement order; however, complement linearisation in the N – bar structures studied exhibited parametric variations. Thus, in some N – bar structures in English the nominal head is ordered to the left whereas in others, the head term occurs to the right. It may therefore be concluded that the principles which determine complement ordering are both Language – Particular and Phrase – specific following insights provided within Phrasal Syntax.Keywords: complement order, complement–head order, head–complement order, language–particular principles
Procedia PDF Downloads 34814915 Current Methods for Drug Property Prediction in the Real World
Authors: Jacob Green, Cecilia Cabrera, Maximilian Jakobs, Andrea Dimitracopoulos, Mark van der Wilk, Ryan Greenhalgh
Abstract:
Predicting drug properties is key in drug discovery to enable de-risking of assets before expensive clinical trials and to find highly active compounds faster. Interest from the machine learning community has led to the release of a variety of benchmark datasets and proposed methods. However, it remains unclear for practitioners which method or approach is most suitable, as different papers benchmark on different datasets and methods, leading to varying conclusions that are not easily compared. Our large-scale empirical study links together numerous earlier works on different datasets and methods, thus offering a comprehensive overview of the existing property classes, datasets, and their interactions with different methods. We emphasise the importance of uncertainty quantification and the time and, therefore, cost of applying these methods in the drug development decision-making cycle. To the best of the author's knowledge, it has been observed that the optimal approach varies depending on the dataset and that engineered features with classical machine learning methods often outperform deep learning. Specifically, QSAR datasets are typically best analysed with classical methods such as Gaussian Processes, while ADMET datasets are sometimes better described by Trees or deep learning methods such as Graph Neural Networks or language models. Our work highlights that practitioners do not yet have a straightforward, black-box procedure to rely on and sets a precedent for creating practitioner-relevant benchmarks. Deep learning approaches must be proven on these benchmarks to become the practical method of choice in drug property prediction.Keywords: activity (QSAR), ADMET, classical methods, drug property prediction, empirical study, machine learning
Procedia PDF Downloads 8114914 Neural Networks-based Acoustic Annoyance Model for Laptop Hard Disk Drive
Authors: Yichao Ma, Chengsiong Chin, Wailok Woo
Abstract:
Since the last decade, there has been a rapid growth in digital multimedia, such as high-resolution media files and three-dimentional movies. Hence, there is a need for large digital storage such as Hard Disk Drive (HDD). As such, users expect to have a quieter HDD in their laptop. In this paper, a jury test has been conducted on a group of 34 people where 17 of them are students who is the potential consumer, and the remaining are engineers who know the HDD. A total 13 HDD sound samples have been selected from over hundred HDD noise recordings. These samples are selected based on an agreed subjective feeling. The samples are played to the participants using head acoustic playback system which enabled them to experience as similar as possible the same environment as have been recorded. Analysis has been conducted and the obtained results have indicated different group has different perception over the noises. Two neural network-based acoustic annoyance models are established based on back propagation neural network. Four psychoacoustic metrics, loudness, sharpness, roughness and fluctuation strength, are used as the input of the model, and the subjective evaluation results are taken as the output. The developed models are reasonably accurate in simulating both training and test samples.Keywords: hdd noise, jury test, neural network model, psychoacoustic annoyance
Procedia PDF Downloads 43814913 A Large-Strain Thermoviscoplastic Damage Model
Authors: João Paulo Pascon
Abstract:
A constitutive model accounting for large strains, thermoviscoplasticity, and ductile damage evolution is proposed in the present work. To this end, a fully Lagrangian framework is employed, considering plane stress conditions and multiplicative split of the deformation gradient. The full model includes Gurson’s void growth, nucleation and coalescence, plastic work heating, strain and strain-rate hardening, thermal softening, and heat conductivity. The contribution of the work is the combination of all the above-mentioned features within the finite-strain setting. The model is implemented in a computer code using triangular finite elements and nonlinear analysis. Two mechanical examples involving ductile damage and finite strain levels are analyzed: an inhomogeneous tension specimen and the necking problem. Results demonstrate the capabilities of the developed formulation regarding ductile fracture and large deformations.Keywords: ductile damage model, finite element method, large strains, thermoviscoplasticity
Procedia PDF Downloads 8614912 Secure Bio Semantic Computing Scheme
Authors: Hiroshi Yamaguchi, Phillip C. Y. Sheu, Ryo Fujita, Shigeo Tsujii
Abstract:
In this paper, the secure BioSemantic Scheme is presented to bridge biological/biomedical research problems and computational solutions via semantic computing. Due to the diversity of problems in various research fields, the semantic capability description language (SCDL) plays and important role as a common language and generic form for problem formalization. SCDL is expected the essential for future semantic and logical computing in Biosemantic field. We show several example to Biomedical problems in this paper. Moreover, in the coming age of cloud computing, the security problem is considered to be crucial issue and we presented a practical scheme to cope with this problem.Keywords: biomedical applications, private information retrieval (PIR), semantic capability description language (SCDL), semantic computing
Procedia PDF Downloads 39014911 Pushing the Boundary of Parallel Tractability for Ontology Materialization via Boolean Circuits
Authors: Zhangquan Zhou, Guilin Qi
Abstract:
Materialization is an important reasoning service for applications built on the Web Ontology Language (OWL). To make materialization efficient in practice, current research focuses on deciding tractability of an ontology language and designing parallel reasoning algorithms. However, some well-known large-scale ontologies, such as YAGO, have been shown to have good performance for parallel reasoning, but they are expressed in ontology languages that are not parallelly tractable, i.e., the reasoning is inherently sequential in the worst case. This motivates us to study the problem of parallel tractability of ontology materialization from a theoretical perspective. That is we aim to identify the ontologies for which materialization is parallelly tractable, i.e., in the NC complexity. Since the NC complexity is defined based on Boolean circuit that is widely used to investigate parallel computing problems, we first transform the problem of materialization to evaluation of Boolean circuits, and then study the problem of parallel tractability based on circuits. In this work, we focus on datalog rewritable ontology languages. We use Boolean circuits to identify two classes of datalog rewritable ontologies (called parallelly tractable classes) such that materialization over them is parallelly tractable. We further investigate the parallel tractability of materialization of a datalog rewritable OWL fragment DHL (Description Horn Logic). Based on the above results, we analyze real-world datasets and show that many ontologies expressed in DHL belong to the parallelly tractable classes.Keywords: ontology materialization, parallel reasoning, datalog, Boolean circuit
Procedia PDF Downloads 271