Search results for: hard segments
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1484

Search results for: hard segments

314 Effect of Plant Density and Planting Pattern on Yield and Quality of Single Cross 704 Silage Corn (Zea mays L.) in Isfahan

Authors: Seyed Mohammad Ali Zahedi

Abstract:

This field experiment was conducted in Isfahan in 2011 in order to study the effect of plant density and planting pattern on growth, yield and quality of silage corn (SC 704) using a randomized complete block design with split plot layout and four replications. The main plot consisted of three planting patterns (60 and 75 cm single planting row and 75 cm double planting row referred to as 60S, 75S and 75T, respectively). The subplots consisted of four levels of plant densities (65000, 80000, 95000 and 110000 plants per hectare). Each subplot consisted of 7 rows, each with 10m length. Vegetative and reproductive characteristics of plants at silking and hard dough stages (when the plants were harvested for silage) were evaluated. Results of variance analysis showed that the effects of planting pattern and plant density were significant on leaf area per plant, leaf area index (at silking), plant height, stem diameter, dry weights of leaf, stem and ear in silking and harvest stages and on fresh and dry yield, dry matter percentage and crude protein percentage at harvest. There was no planting pattern × plant density interaction for these parameters. As row space increased from 60 cm with single planting to 75 cm with single planting, leaf area index and plant height increased, but leaf area per plant, stem diameter, dry weight of leaf, stem and ear, dry matter percentage, dry matter yield and crude protein percentage decreased. Dry matter yield reduced from 24.9 to 18.5 t/ha and crude protein percentage decreased from 6.11 to 5.60 percent. When the plant density increased from 65000 to 110000 plant per hectare, leaf area index, plant height, dry weight of leaf, stem and ear and dry matter yield increased from 19.2 to 23.3 t/ha, whereas leaf area per plant, stem diameter, dry matter percentage and crude protein percentage decreased from 6.30 to 5.25. The best results were obtained with 60 cm row distance with single planting and 110000 plants per hectare.

Keywords: silage corn, plant density, planting pattern, yield

Procedia PDF Downloads 333
313 Application of Recycled Tungsten Carbide Powder for Fabrication of Iron Based Powder Metallurgy Alloy

Authors: Yukinori Taniguchi, Kazuyoshi Kurita, Kohei Mizuta, Keigo Nishitani, Ryuichi Fukuda

Abstract:

Tungsten carbide is widely used as a tool material in metal manufacturing process. Since tungsten is typical rare metal, establishment of recycle process of tungsten carbide tools and restore into cemented carbide material bring great impact to metal manufacturing industry. Recently, recycle process of tungsten carbide has been developed and established gradually. However, the demands for quality of cemented carbide tool are quite severe because hardness, toughness, anti-wear ability, heat resistance, fatigue strength and so on should be guaranteed for precision machining and tool life. Currently, it is hard to restore the recycled tungsten carbide powder entirely as raw material for new processed cemented carbide tool. In this study, to suggest positive use of recycled tungsten carbide powder, we have tried to fabricate a carbon based sintered steel which shows reinforced mechanical properties with recycled tungsten carbide powder. We have made set of newly designed sintered steels. Compression test of sintered specimen in density ratio of 0.85 (which means 15% porosity inside) has been conducted. As results, at least 1.7 times higher in nominal strength in the amount of 7.0 wt.% was shown in recycled WC powder. The strength reached to over 600 MPa for the Fe-WC-Co-Cu sintered alloy. Wear test has been conducted by using ball-on-disk type friction tester using 5 mm diameter ball with normal force of 2 N in the dry conditions. Wear amount after 1,000 m running distance shows that about 1.5 times longer life was shown in designed sintered alloy. Since results of tensile test showed that same tendency in previous testing, it is concluded that designed sintered alloy can be used for several mechanical parts with special strength and anti-wear ability in relatively low cost due to recycled tungsten carbide powder.

Keywords: tungsten carbide, recycle process, compression test, powder metallurgy, anti-wear ability

Procedia PDF Downloads 245
312 Powered Two-Wheeler Rider’s Comfort over Road Sections with Skew Superelevation

Authors: Panagiotis Lemonakis, Nikolaos Moisiadis, Andromachi Gkoutzini, George Kaliabetsos, Nikos Eliou

Abstract:

The proper surface water drainage not only affects vehicle movement dynamics but also increases the likelihood of an accident due to the fact that inadequate drainage is associated with potential hydroplaning and splash and spray driving conditions. Nine solutions have been proposed to address hydroplaning in sections with inadequate drainage, e.g., augmented superelevation and longitudinal rates, reduction of runoff length, and skew superelevation. The latter has been extensively implemented in highways recently, enhancing the safety level in the applied road segments in regards to the effective drainage of the rainwater. However, the concept of the skew superelevation has raised concerns regarding the driver’s comfort when traveling over skew superelevation sections, particularly at high speeds. These concerns alleviated through the concept of the round-up skew superelevation, which reduces both the lateral and the vertical acceleration imposed to the drivers and hence, improves comfort and traffic safety. Various research studies aimed at investigating driving comfort by evaluating the lateral and vertical accelerations sustained by the road users and vehicles. These studies focused on the influence of the skew superelevation to passenger cars, buses and trucks, and the drivers themselves, traveling at a certain range of speeds either below or above the design speed. The outcome of these investigations which based on the use of simulations, revealed that the imposed accelerations did not exceed the statutory thresholds even when the travelling speed was significantly greater than the design speed. Nevertheless, the effect of the skew superelevation to other vehicle types for instance, motorcycles, has not been investigated so far. The present research study aims to bridge this gap by investigating the impact of skew superelevation on the motorcycle rider’s comfort. Power two-wheeler riders are susceptible to any changes on the pavement surface and therefore a comparison between the traditional superelevation practice and the skew superelevation concept is of paramount importance. The methodology based on the utilization of sophisticated software in order to design the model of the road for several values of the longitudinal slope. Based on the values of the slopes and the use of a mathematical equation, the accelerations imposed on the wheel of the motorcycle were calculated. Due to the fact that the final aim of the study is the influence of the skew superelevation to the rider, it was deemed necessary to convey the calculated accelerations from the wheel to the rider. That was accomplished by implementing the quarter car suspension model adjusted to the features of two-wheeler vehicles. Finally, the accelerations derived from this process evaluated according to specific thresholds originated from the International Organization for Standardization, which correspond to certain levels of comfort. The most important conclusion drawn is that the comfort of the riders is not dependent on the form of road gradient to a great extent due to the fact that the vertical acceleration imposed to the riders took similar values regardless of the value of the longitudinal slope.

Keywords: acceleration, comfort, motorcycle, safety, skew superelevation

Procedia PDF Downloads 149
311 Understanding the Productivity Effect on Industrial Management: The Portuguese Wood Furniture Industry Case Study

Authors: Jonas A. R. H. Lima, Maria Antonia Carravilla

Abstract:

As productivity concepts are widely related to industrial savings, it is becoming particularly important in a more and more competitive world, to really understand how productivity can be well used in industrial management techniques. Nowadays, consumers are no more willing to pay for mistakes and inefficiencies. Therefore, one way for companies to stay competitive is to control and increase their productivity. This study aims to define clearly the productivity concept, understand how a company can affect productivity, and, if possible, identify the relation between each identified productivity factor. This will help managers, by clarifying the main issues behind productivity concepts and proposing a methodology to measure, control and increase productivity. The main questions to be answered are: what is the importance of productivity for the Portuguese Wood Furniture Industry? Is it possible to control productivity internally, or is it a phenomenon external to companies, hard or even impossible to control? How to understand, control and adjust productivity performance? How to make productivity to become one main asset for maximizing the use of the available resources? This essay will follow a constructive approach mostly based in the research hypothesis mentioned above. For that, a literature review is being done to find the main conceptual frameworks and empirical studies that already exist, and by doing so, highlight eventual knowledge or conflicting research to be addressed in this work. We expect to build theoretical explanations and test theoretical predictions from participants understandings and own experiences, by elaborating field surveys and interviews, to select adjusted productivity indicators and analyze the productivity evolution according the adjustments on other variables. Its intended the conduction of an exploratory work that can simultaneous clarify productivity concepts, objectives, and define frameworks. This investigation intends to migrate from merely academic concepts to a daily basis operational reality of the companies from the Portuguese Wood Furniture Industry highlighting productivity increased importance within modern engineering and industrial management. The ambition is to clarify, systemize and develop a management tool that may not only control but positively influence the way resources are used.

Keywords: industrial management, motivation, productivity, performance indicators, reward management, wood furniture industry

Procedia PDF Downloads 225
310 Sparse Representation Based Spatiotemporal Fusion Employing Additional Image Pairs to Improve Dictionary Training

Authors: Dacheng Li, Bo Huang, Qinjin Han, Ming Li

Abstract:

Remotely sensed imagery with the high spatial and temporal characteristics, which it is hard to acquire under the current land observation satellites, has been considered as a key factor for monitoring environmental changes over both global and local scales. On a basis of the limited high spatial-resolution observations, challenged studies called spatiotemporal fusion have been developed for generating high spatiotemporal images through employing other auxiliary low spatial-resolution data while with high-frequency observations. However, a majority of spatiotemporal fusion approaches yield to satisfactory assumption, empirical but unstable parameters, low accuracy or inefficient performance. Although the spatiotemporal fusion methodology via sparse representation theory has advantage in capturing reflectance changes, stability and execution efficiency (even more efficient when overcomplete dictionaries have been pre-trained), the retrieval of high-accuracy dictionary and its response to fusion results are still pending issues. In this paper, we employ additional image pairs (here each image-pair includes a Landsat Operational Land Imager and a Moderate Resolution Imaging Spectroradiometer acquisitions covering the partial area of Baotou, China) only into the coupled dictionary training process based on K-SVD (K-means Singular Value Decomposition) algorithm, and attempt to improve the fusion results of two existing sparse representation based fusion models (respectively utilizing one and two available image-pair). The results show that more eligible image pairs are probably related to a more accurate overcomplete dictionary, which generally indicates a better image representation, and is then contribute to an effective fusion performance in case that the added image-pair has similar seasonal aspects and image spatial structure features to the original image-pair. It is, therefore, reasonable to construct multi-dictionary training pattern for generating a series of high spatial resolution images based on limited acquisitions.

Keywords: spatiotemporal fusion, sparse representation, K-SVD algorithm, dictionary learning

Procedia PDF Downloads 254
309 Using Multiomic Plasma Profiling From Liquid Biopsies to Identify Potential Signatures for Disease Diagnostics in Late-Stage Non-small Cell Lung Cancer (NSCLC) in Trinidad and Tobago

Authors: Nicole Ramlachan, Samuel Mark West

Abstract:

Lung cancer is the leading cause of cancer-associated deaths in North America, with the vast majority being non-small cell lung cancer (NSCLC), with a five-year survival rate of only 24%. Non-invasive discovery of biomarkers associated with early-diagnosis of NSCLC can enable precision oncology efforts using liquid biopsy-based multiomics profiling of plasma. Although tissue biopsies are currently the gold standard for tumor profiling, this method presents many limitations since these are invasive, risky, and sometimes hard to obtain as well as only giving a limited tumor profile. Blood-based tests provides a less-invasive, more robust approach to interrogate both tumor- and non-tumor-derived signals. We intend to examine 30 stage III-IV NSCLC patients pre-surgery and collect plasma samples.Cell-free DNA (cfDNA) will be extracted from plasma, and next-generation sequencing (NGS) performed. Through the analysis of tumor-specific alterations, including single nucleotide variants (SNVs), insertions, deletions, copy number variations (CNVs), and methylation alterations, we intend to identify tumor-derived DNA—ctDNA among the total pool of cfDNA. This would generate data to be used as an accurate form of cancer genotyping for diagnostic purposes. Using liquid biopsies offer opportunities to improve the surveillance of cancer patients during treatment and would supplement current diagnosis and tumor profiling strategies previously not readily available in Trinidad and Tobago. It would be useful and advantageous to use this in diagnosis and tumour profiling as well as to monitor cancer patients, providing early information regarding disease evolution and treatment efficacy, and reorient treatment strategies in, timethereby improving clinical oncology outcomes.

Keywords: genomics, multiomics, clinical genetics, genotyping, oncology, diagnostics

Procedia PDF Downloads 156
308 The Integration of Geographical Information Systems and Capacitated Vehicle Routing Problem with Simulated Demand for Humanitarian Logistics in Tsunami-Prone Area: A Case Study of Phuket, Thailand

Authors: Kiatkulchai Jitt-Aer, Graham Wall, Dylan Jones

Abstract:

As a result of the Indian Ocean tsunami in 2004, logistics applied to disaster relief operations has received great attention in the humanitarian sector. As learned from such disaster, preparing and responding to the aspect of delivering essential items from distribution centres to affected locations are of the importance for relief operations as the nature of disasters is uncertain especially in suffering figures, which are normally proportional to quantity of supplies. Thus, this study proposes a spatial decision support system (SDSS) for humanitarian logistics by integrating Geographical Information Systems (GIS) and the capacitated vehicle routing problem (CVRP). The GIS is utilised for acquiring demands simulated from the tsunami flooding model of the affected area in the first stage, and visualising the simulation solutions in the last stage. While CVRP in this study encompasses designing the relief routes of a set of homogeneous vehicles from a relief centre to a set of geographically distributed evacuation points in which their demands are estimated by using both simulation and randomisation techniques. The CVRP is modeled as a multi-objective optimization problem where both total travelling distance and total transport resources used are minimized, while demand-cost efficiency of each route is maximized in order to determine route priority. As the model is a NP-hard combinatorial optimization problem, the Clarke and Wright Saving heuristics is proposed to solve the problem for the near-optimal solutions. The real-case instances in the coastal area of Phuket, Thailand are studied to perform the SDSS that allows a decision maker to visually analyse the simulation scenarios through different decision factors.

Keywords: demand simulation, humanitarian logistics, geographical information systems, relief operations, capacitated vehicle routing problem

Procedia PDF Downloads 243
307 Self-Assembled Laser-Activated Plasmonic Substrates for High-Throughput, High-Efficiency Intracellular Delivery

Authors: Marinna Madrid, Nabiha Saklayen, Marinus Huber, Nicolas Vogel, Christos Boutopoulos, Michel Meunier, Eric Mazur

Abstract:

Delivering material into cells is important for a diverse range of biological applications, including gene therapy, cellular engineering and imaging. We present a plasmonic substrate for delivering membrane-impermeable material into cells at high throughput and high efficiency while maintaining cell viability. The substrate fabrication is based on an affordable and fast colloidal self-assembly process. When illuminated with a femtosecond laser, the light interacts with the electrons at the surface of the metal substrate, creating localized surface plasmons that form bubbles via energy dissipation in the surrounding medium. These bubbles come into close contact with the cell membrane to form transient pores and enable entry of membrane-impermeable material via diffusion. We use fluorescence microscopy and flow cytometry to verify delivery of membrane-impermeable material into HeLa CCL-2 cells. We show delivery efficiency and cell viability data for a range of membrane-impermeable cargo, including dyes and biologically relevant material such as siRNA. We estimate the effective pore size by determining delivery efficiency for hard fluorescent spheres with diameters ranging from 20 nm to 2 um. To provide insight to the cell poration mechanism, we relate the poration data to pump-probe measurements of micro- and nano-bubble formation on the plasmonic substrate. Finally, we investigate substrate stability and reusability by using scanning electron microscopy (SEM) to inspect for damage on the substrate after laser treatment. SEM images show no visible damage. Our findings indicate that self-assembled plasmonic substrates are an affordable tool for high-throughput, high-efficiency delivery of material into mammalian cells.

Keywords: femtosecond laser, intracellular delivery, plasmonic, self-assembly

Procedia PDF Downloads 525
306 Searching for the ‘Why’ of Gendered News: Journalism Practices and Societal Contexts

Authors: R. Simões, M. Silveirinha

Abstract:

Driven by the need to understand the results of previous research that clearly shows deep unbalances of the media discourses about women and men in spite of the growing numbers of female journalists, our paper aims to progress from the 'what' to the 'why' of these unbalanced representations. Furthermore, it does so at a time when journalism is undergoing a dramatic change in terms of professional practices and in how media organizations are organized and run, affecting women in particular. While some feminist research points to the fact that female and male journalists evaluate the role of the news and production methods in similar ways feminist theorizing also suggests that thought and knowledge are highly influenced by social identity, which is also inherently affected by the experiences of gender. This is particularly important at a time of deep societal and professional changes. While there are persuasive discussions of gender identities at work in newsrooms in various countries studies on the issue will benefit from cases that focus on the particularities of local contexts. In our paper, we present one such case: the case of Portugal, a country hit hard by austerity measures that have affected all cultural industries including journalism organizations, already feeling the broader impacts of the larger societal changes of the media landscape. Can we gender these changes? How are they felt and understood by female and male journalists? And how are these discourses framed by androcentric, feminist and post-feminist sensibilities? Foregrounding questions of gender, our paper seeks to explore some of the interactions of societal and professional forces, identifying their gendered character and outlining how they shape journalism work in general and the production of unbalanced gender representations in particular. We do so grounded in feminist studies of journalism as well as feminist organizational and work studies, looking at a corpus of 20 in-depth interviews of female and male Portuguese journalists. The research findings illustrate how gender in journalism practices interacts with broader experiences of the cultural and economic contexts and show the ambivalences of these interactions in news organizations.

Keywords: gender, journalism, newsroom culture, Portuguese journalists

Procedia PDF Downloads 394
305 Image-Based UAV Vertical Distance and Velocity Estimation Algorithm during the Vertical Landing Phase Using Low-Resolution Images

Authors: Seyed-Yaser Nabavi-Chashmi, Davood Asadi, Karim Ahmadi, Eren Demir

Abstract:

The landing phase of a UAV is very critical as there are many uncertainties in this phase, which can easily entail a hard landing or even a crash. In this paper, the estimation of relative distance and velocity to the ground, as one of the most important processes during the landing phase, is studied. Using accurate measurement sensors as an alternative approach can be very expensive for sensors like LIDAR, or with a limited operational range, for sensors like ultrasonic sensors. Additionally, absolute positioning systems like GPS or IMU cannot provide distance to the ground independently. The focus of this paper is to determine whether we can measure the relative distance and velocity of UAV and ground in the landing phase using just low-resolution images taken by a monocular camera. The Lucas-Konda feature detection technique is employed to extract the most suitable feature in a series of images taken during the UAV landing. Two different approaches based on Extended Kalman Filters (EKF) have been proposed, and their performance in estimation of the relative distance and velocity are compared. The first approach uses the kinematics of the UAV as the process and the calculated optical flow as the measurement; On the other hand, the second approach uses the feature’s projection on the camera plane (pixel position) as the measurement while employing both the kinematics of the UAV and the dynamics of variation of projected point as the process to estimate both relative distance and relative velocity. To verify the results, a sequence of low-quality images taken by a camera that is moving on a specifically developed testbed has been used to compare the performance of the proposed algorithm. The case studies show that the quality of images results in considerable noise, which reduces the performance of the first approach. On the other hand, using the projected feature position is much less sensitive to the noise and estimates the distance and velocity with relatively high accuracy. This approach also can be used to predict the future projected feature position, which can drastically decrease the computational workload, as an important criterion for real-time applications.

Keywords: altitude estimation, drone, image processing, trajectory planning

Procedia PDF Downloads 106
304 Measurement and Simulation of Axial Neutron Flux Distribution in Dry Tube of KAMINI Reactor

Authors: Manish Chand, Subhrojit Bagchi, R. Kumar

Abstract:

A new dry tube (DT) has been installed in the tank of KAMINI research reactor, Kalpakkam India. This tube will be used for neutron activation analysis of small to large samples and testing of neutron detectors. DT tube is 375 cm height and 7.5 cm in diameter, located 35 cm away from the core centre. The experimental thermal flux at various axial positions inside the tube has been measured by irradiating the flux monitor (¹⁹⁷Au) at 20kW reactor power. The measured activity of ¹⁹⁸Au and the thermal cross section of ¹⁹⁷Au (n,γ) ¹⁹⁸Au reaction were used for experimental thermal flux measurement. The flux inside the tube varies from 10⁹ to 10¹⁰ and maximum flux was (1.02 ± 0.023) x10¹⁰ n cm⁻²s⁻¹ at 36 cm from the bottom of the tube. The Au and Zr foils without and with cadmium cover of 1-mm thickness were irradiated at the maximum flux position in the DT to find out the irradiation specific input parameters like sub-cadmium to epithermal neutron flux ratio (f) and the epithermal neutron flux shape factor (α). The f value was 143 ± 5, indicates about 99.3% thermal neutron component and α value was -0.2886 ± 0.0125, indicates hard epithermal neutron spectrum due to insufficient moderation. The measured flux profile has been validated using theoretical model of KAMINI reactor through Monte Carlo N-Particle Code (MCNP). In MCNP, the complex geometry of the entire reactor is modelled in 3D, ensuring minimum approximations for all the components. Continuous energy cross-section data from ENDF-B/VII.1 as well as S (α, β) thermal neutron scattering functions are considered. The neutron flux has been estimated at the corresponding axial locations of the DT using mesh tally. The thermal flux obtained from the experiment shows good agreement with the theoretically predicted values by MCNP, it was within ± 10%. It can be concluded that this MCNP model can be utilized for calculating other important parameters like neutron spectra, dose rate, etc. and multi elemental analysis can be carried out by irradiating the sample at maximum flux position using measured f and α parameters by k₀-NAA standardization.

Keywords: neutron flux, neutron activation analysis, neutron flux shape factor, MCNP, Monte Carlo N-Particle Code

Procedia PDF Downloads 152
303 Field Study of Chlorinated Aliphatic Hydrocarbons Degradation in Contaminated Groundwater via Micron Zero-Valent Iron Coupled with Biostimulation

Authors: Naijin Wu, Peizhong Li, Haijian Wang, Wenxia Wei, Yun Song

Abstract:

Chlorinated aliphatic hydrocarbons (CAHs) pollution poses a severe threat to human health and is persistent in groundwater. Although chemical reduction or bioremediation is effective, it is still hard to achieve their complete and rapid dechlorination. Recently, the combination of zero-valent iron and biostimulation has been considered to be one of the most promising strategies, but field studies of this technology are scarce. In a typical site contaminated by various types of CAHs, basic physicochemical parameters of groundwater, CAHs and their product concentrations, and microbial abundance and diversity were monitored after a remediation slurry containing both micron zero-valent iron (mZVI) and biostimulation components were directly injected into the aquifer. Results showed that groundwater could form and keep low oxidation-reduction potential (ORP), a neutral pH, and anoxic conditions after different degrees of fluctuations, which was benefit for the reductive dechlorination of CAHs. The injection also caused an obvious increase in the total organic carbon (TOC) concentration and sulfate reduction. After 253 days post-injection, the mean concentration of total chlorinated ethylene (CEE) from two monitoring wells decreased from 304 μg/L to 8 μg/L, and total chlorinated ethane (CEA) decreased from 548 μg/L to 108 μg/L. Occurrence of chloroethane (CA) suggested that hydrogenolysis dechlorination was one of the main degradation pathways for CEA, and also hints that biological dechlorination was activated. A significant increase of ethylene at day 67 post-injection indicated that dechlorination was complete. Additionally, the total bacterial counts increased by 2-3 orders of magnitude after 253 days post-injection. And the microbial species richness decreased and gradually changed to anaerobic/fermentative bacteria. The relative abundance of potential degradation bacteria increased corresponding to the degradation of CAHs. This work demonstrates that mZVI and biostimulation can be combined to achieve the efficient removal of various CAHs from contaminated groundwater sources.

Keywords: chlorinated aliphatic hydrocarbons, groundwater, field study, zero-valent iron, biostimulation

Procedia PDF Downloads 159
302 On Cloud Computing: A Review of the Features

Authors: Assem Abdel Hamed Mousa

Abstract:

The Internet of Things probably already influences your life. And if it doesn’t, it soon will, say computer scientists; Ubiquitous computing names the third wave in computing, just now beginning. First were mainframes, each shared by lots of people. Now we are in the personal computing era, person and machine staring uneasily at each other across the desktop. Next comes ubiquitous computing, or the age of calm technology, when technology recedes into the background of our lives. Alan Kay of Apple calls this "Third Paradigm" computing. Ubiquitous computing is essentially the term for human interaction with computers in virtually everything. Ubiquitous computing is roughly the opposite of virtual reality. Where virtual reality puts people inside a computer-generated world, ubiquitous computing forces the computer to live out here in the world with people. Virtual reality is primarily a horse power problem; ubiquitous computing is a very difficult integration of human factors, computer science, engineering, and social sciences. The approach: Activate the world. Provide hundreds of wireless computing devices per person per office, of all scales (from 1" displays to wall sized). This has required new work in operating systems, user interfaces, networks, wireless, displays, and many other areas. We call our work "ubiquitous computing". This is different from PDA's, dynabooks, or information at your fingertips. It is invisible; everywhere computing that does not live on a personal device of any sort, but is in the woodwork everywhere. The initial incarnation of ubiquitous computing was in the form of "tabs", "pads", and "boards" built at Xerox PARC, 1988-1994. Several papers describe this work, and there are web pages for the Tabs and for the Boards (which are a commercial product now): Ubiquitous computing will drastically reduce the cost of digital devices and tasks for the average consumer. With labor intensive components such as processors and hard drives stored in the remote data centers powering the cloud , and with pooled resources giving individual consumers the benefits of economies of scale, monthly fees similar to a cable bill for services that feed into a consumer’s phone.

Keywords: internet, cloud computing, ubiquitous computing, big data

Procedia PDF Downloads 377
301 Documenting the 15th Century Prints with RTI

Authors: Peter Fornaro, Lothar Schmitt

Abstract:

The Digital Humanities Lab and the Institute of Art History at the University of Basel are collaborating in the SNSF research project ‘Digital Materiality’. Its goal is to develop and enhance existing methods for the digital reproduction of cultural heritage objects in order to support art historical research. One part of the project focuses on the visualization of a small eye-catching group of early prints that are noteworthy for their subtle reliefs and glossy surfaces. Additionally, this group of objects – known as ‘paste prints’ – is characterized by its fragile state of preservation. Because of the brittle substances that were used for their production, most paste prints are heavily damaged and thus very hard to examine. These specific material properties make a photographic reproduction extremely difficult. To obtain better results we are working with Reflectance Transformation Imaging (RTI), a computational photographic method that is already used in archaeological and cultural heritage research. This technique allows documenting how three-dimensional surfaces respond to changing lighting situations. Our first results show that RTI can capture the material properties of paste prints and their current state of preservation more accurately than conventional photographs, although there are limitations with glossy surfaces because the mathematical models that are included in RTI are kept simple in order to keep the software robust and easy to use. To improve the method, we are currently developing tools for a more detailed analysis and simulation of the reflectance behavior. An enhanced analytical model for the representation and visualization of gloss will increase the significance of digital representations of cultural heritage objects. For collaborative efforts, we are working on a web-based viewer application for RTI images based on WebGL in order to make acquired data accessible to a broader international research community. At the ICDH Conference, we would like to present unpublished results of our work and discuss the implications of our concept for art history, computational photography and heritage science.

Keywords: art history, computational photography, paste prints, reflectance transformation imaging

Procedia PDF Downloads 273
300 Teaching Non-Euclidean Geometries to Learn Euclidean One: An Experimental Study

Authors: Silvia Benvenuti, Alessandra Cardinali

Abstract:

In recent years, for instance, in relation to the Covid 19 pandemic and the evidence of climate change, it is becoming quite clear that the development of a young kid into an adult citizen requires a solid scientific background. Citizens are required to exert logical thinking and know the methods of science in order to adapt, understand, and develop as persons. Mathematics sits at the core of these required skills: learning the axiomatic method is fundamental to understand how hard sciences work and helps in consolidating logical thinking, which will be useful for the entire life of a student. At the same time, research shows that the axiomatic study of geometry is a problematic topic for students, even for those with interest in mathematics. With this in mind, the main goals of the research work we will describe are: (1) to show whether non-Euclidean geometries can be a tool to allow students to consolidate the knowledge of Euclidean geometries by developing it in a critical way; (2) to promote the understanding of the modern axiomatic method in geometry; (3) to give students a new perspective on mathematics so that they can see it as a creative activity and a widely discussed topic with a historical background. One of the main issues related to the state-of-the-art in this topic is the shortage of experimental studies with students. For this reason, our aim is to show further experimental evidence of the potential benefits of teaching non-Euclidean geometries at high school, based on data collected from a study started in 2005 in the frame of the Italian National Piano Lauree Scientifiche, continued by a teacher training organized in September 2018, perfected in a pilot study that involved 77 high school students during the school years 2018-2019 and 2019-2020. and finally implemented through an experimental study conducted in 2020-21 with 87 high school students. Our study shows that there is potential for further research to challenge current conceptions of the school mathematics curriculum and of the capabilities of high school mathematics students.

Keywords: Non-Euclidean geometries, beliefs about mathematics, questionnaires, modern axiomatic method

Procedia PDF Downloads 71
299 Soybean Oil Based Phase Change Material for Thermal Energy Storage

Authors: Emre Basturk, Memet Vezir Kahraman

Abstract:

In many developing countries, with the rapid economic improvements, energy shortage and environmental issues have become a serious problem. Therefore, it has become a very critical issue to improve energy usage efficiency and also protect the environment. Thermal energy storage system is an essential approach to match the thermal energy claim and supply. Thermal energy can be stored by heating, cooling or melting a material with the energy and then enhancing accessible when the procedure is reversed. The overall thermal energy storage techniques are sorted as; latent heat or sensible heat thermal energy storage technology segments. Among these methods, latent heat storage is the most effective method of collecting thermal energy. Latent heat thermal energy storage depend on the storage material, emitting or discharging heat as it undergoes a solid to liquid, solid to solid or liquid to gas phase change or vice versa. Phase change materials (PCMs) are promising materials for latent heat storage applications due to their capacities to accumulate high latent heat storage per unit volume by phase change at an almost constant temperature. Phase change materials (PCMs) are being utilized to absorb, collect and discharge thermal energy during the cycle of melting and freezing, converting from one phase to another. Phase Change Materials (PCMs) can generally be arranged into three classes: organic materials, salt hydrates and eutectics. Many kinds of organic and inorganic PCMs and their blends have been examined as latent heat storage materials. Organic PCMs are rather expensive and they have average latent heat storage per unit volume and also have low density. Most organic PCMs are combustible in nature and also have a wide range of melting point. Organic PCMs can be categorized into two major categories: non-paraffinic and paraffin materials. Paraffin materials have been extensively used, due to their high latent heat and right thermal characteristics, such as minimal super cooling, varying phase change temperature, low vapor pressure while melting, good chemical and thermal stability, and self-nucleating behavior. Ultraviolet (UV)-curing technology has been generally used because it has many advantages, such as low energy consumption , high speed, high chemical stability, room-temperature operation, low processing costs and environmental friendly. For many years, PCMs have been used for heating and cooling industrial applications including textiles, refrigerators, construction, transportation packaging for temperature-sensitive products, a few solar energy based systems, biomedical and electronic materials. In this study, UV-curable, fatty alcohol containing soybean oil based phase change materials (PCMs) were obtained and characterized. The phase transition behaviors and thermal stability of the prepared UV-cured biobased PCMs were analyzed by differential scanning calorimetry (DSC) and thermogravimetric analysis (TGA). The heating process phase change enthalpy is measured between 30 and 68 J/g, and the freezing process phase change enthalpy is found between 18 and 70 J/g. The decomposition of UVcured PCMs started at 260 ºC and reached a maximum of 430 ºC.

Keywords: fatty alcohol, phase change material, thermal energy storage, UV curing

Procedia PDF Downloads 374
298 The Roles of Mandarin and Local Dialect in the Acquisition of L2 English Consonants Among Chinese Learners of English: Evidence From Suzhou Dialect Areas

Authors: Weijing Zhou, Yuting Lei, Francis Nolan

Abstract:

In the domain of second language acquisition, whenever pronunciation errors or acquisition difficulties are found, researchers habitually attribute them to the negative transfer of the native language or local dialect. To what extent do Mandarin and local dialects affect English phonological acquisition for Chinese learners of English as a foreign language (EFL)? Little evidence, however, has been found via empirical research in China. To address this core issue, the present study conducted phonetic experiments to explore the roles of local dialects and Mandarin in Chinese EFL learners’ acquisition of L2 English consonants. Besides Mandarin, the sole national language in China, Suzhou dialect was selected as the target local dialect because of its distinct phonology from Mandarin. The experimental group consisted of 30 junior English majors at Yangzhou University, who were born and lived in Suzhou, acquired Suzhou Dialect since their early childhood, and were able to communicate freely and fluently with each other in Suzhou Dialect, Mandarin as well as English. The consonantal target segments were all the consonants of English, Mandarin and Suzhou Dialect in typical carrier words embedded in the carrier sentence Say again. The control group consisted of two Suzhou Dialect experts, two Mandarin radio broadcasters, and two British RP phoneticians, who served as the standard speakers of the three languages. The reading corpus was recorded and sampled in the phonetic laboratories at Yangzhou University, Soochow University and Cambridge University, respectively, then transcribed, segmented and analyzed acoustically via Praat software, and finally analyzed statistically via EXCEL and SPSS software. The main findings are as follows: First, in terms of correct acquisition rates (CARs) of all the consonants, Mandarin ranked top (92.83%), English second (74.81%) and Suzhou Dialect last (70.35%), and significant differences were found only between the CARs of Mandarin and English and between the CARs of Mandarin and Suzhou Dialect, demonstrating Mandarin was overwhelmingly more robust than English or Suzhou Dialect in subjects’ multilingual phonological ecology. Second, in terms of typical acoustic features, the average duration of all the consonants plus the voice onset time (VOT) of plosives, fricatives, and affricatives in 3 languages were much longer than those of standard speakers; the intensities of English fricatives and affricatives were higher than RP speakers but lower than Mandarin and Suzhou Dialect standard speakers; the formants of English nasals and approximants were significantly different from those of Mandarin and Suzhou Dialects, illustrating the inconsistent acoustic variations between the 3 languages. Thirdly, in terms of typical pronunciation variations or errors, there were significant interlingual interactions between the 3 consonant systems, in which Mandarin consonants were absolutely dominant, accounting for the strong transfer from L1 Mandarin to L2 English instead of from earlier-acquired L1 local dialect to L2 English. This is largely because the subjects were knowingly exposed to Mandarin since their nursery and were strictly required to speak in Mandarin through all the formal education periods from primary school to university.

Keywords: acquisition of L2 English consonants, role of Mandarin, role of local dialect, Chinese EFL learners from Suzhou Dialect areas

Procedia PDF Downloads 92
297 Discerning Divergent Nodes in Social Networks

Authors: Mehran Asadi, Afrand Agah

Abstract:

In data mining, partitioning is used as a fundamental tool for classification. With the help of partitioning, we study the structure of data, which allows us to envision decision rules, which can be applied to classification trees. In this research, we used online social network dataset and all of its attributes (e.g., Node features, labels, etc.) to determine what constitutes an above average chance of being a divergent node. We used the R statistical computing language to conduct the analyses in this report. The data were found on the UC Irvine Machine Learning Repository. This research introduces the basic concepts of classification in online social networks. In this work, we utilize overfitting and describe different approaches for evaluation and performance comparison of different classification methods. In classification, the main objective is to categorize different items and assign them into different groups based on their properties and similarities. In data mining, recursive partitioning is being utilized to probe the structure of a data set, which allow us to envision decision rules and apply them to classify data into several groups. Estimating densities is hard, especially in high dimensions, with limited data. Of course, we do not know the densities, but we could estimate them using classical techniques. First, we calculated the correlation matrix of the dataset to see if any predictors are highly correlated with one another. By calculating the correlation coefficients for the predictor variables, we see that density is strongly correlated with transitivity. We initialized a data frame to easily compare the quality of the result classification methods and utilized decision trees (with k-fold cross validation to prune the tree). The method performed on this dataset is decision trees. Decision tree is a non-parametric classification method, which uses a set of rules to predict that each observation belongs to the most commonly occurring class label of the training data. Our method aggregates many decision trees to create an optimized model that is not susceptible to overfitting. When using a decision tree, however, it is important to use cross-validation to prune the tree in order to narrow it down to the most important variables.

Keywords: online social networks, data mining, social cloud computing, interaction and collaboration

Procedia PDF Downloads 150
296 The Impact of Shifting Trading Pattern from Long-Haul to Short-Sea to the Car Carriers’ Freight Revenues

Authors: Tianyu Wang, Nikita Karandikar

Abstract:

The uncertainty around cost, safety, and feasibility of the decarbonized shipping fuels has made it increasingly complex for the shipping companies to set pricing strategies and forecast their freight revenues going forward. The increase in the green fuel surcharges will ultimately influence the automobile’s consumer prices. The auto shipping demand (ton-miles) has been gradually shifting from long-haul to short-sea trade over the past years following the relocation of the original equipment manufacturer (OEM) manufacturing to regions such as South America and Southeast Asia. The objective of this paper is twofold: 1) to investigate the car-carriers freight revenue development over the years when the trade pattern is gradually shifting towards short-sea exports 2) to empirically identify the quantitative impact of such trade pattern shifting to mainly freight rate, but also vessel size, fleet size as well as Green House Gas (GHG) emission in Roll on-Roll Off (Ro-Ro) shipping. In this paper, a model of analyzing and forecasting ton-miles and freight revenues for the trade routes of AS-NA (Asia to North America), EU-NA (Europe to North America), and SA-NA (South America to North America) is established by deploying Automatic Identification System (AIS) data and the financial results of a selected car carrier company. More specifically, Wallenius Wilhelmsen Logistics (WALWIL), the Norwegian Ro-Ro carrier listed on Oslo Stock Exchange, is selected as the case study company in this paper. AIS-based ton-mile datasets of WALWIL vessels that are sailing into North America region from three different origins (Asia, Europe, and South America), together with WALWIL’s quarterly freight revenues as reported in trade segments, will be investigated and compared for the past five years (2018-2022). Furthermore, ordinary‐least‐square (OLS) regression is utilized to construct the ton-mile demand and freight revenue forecasting. The determinants of trade pattern shifting, such as import tariffs following the China-US trade war and fuel prices following the 0.1% Emission Control Areas (ECA) zone requirement after IMO2020 will be set as key variable inputs to the machine learning model. The model will be tested on another newly listed Norwegian Car Carrier, Hoegh Autoliner, to forecast its 2022 financial results and to validate the accuracy based on its actual results. GHG emissions on the three routes will be compared and discussed based on a constant emission per mile assumption and voyage distances. Our findings will provide important insights about 1) the trade-off evaluation between revenue reduction and energy saving with the new ton-mile pattern and 2) how the trade flow shifting would influence the future need for the vessel and fleet size.

Keywords: AIS, automobile exports, maritime big data, trade flows

Procedia PDF Downloads 118
295 Disrupting Traditional Industries: A Scenario-Based Experiment on How Blockchain-Enabled Trust and Transparency Transform Nonprofit Organizations

Authors: Michael Mertel, Lars Friedrich, Kai-Ingo Voigt

Abstract:

Based on principle-agent theory, an information asymmetry exists in the traditional donation process. Consumers cannot comprehend whether nonprofit organizations (NPOs) use raised funds according to the designated cause after the transaction took place (hidden action). Therefore, charity organizations have tried to appear transparent and gain trust by using the same marketing instruments for decades (e.g., releasing project success reports). However, none of these measures can guarantee consumers that charities will use their donations for the purpose. With awareness of misuse of donations rising due to the Ukraine conflict (e.g., funding crime), consumers are increasingly concerned about the destination of their charitable purposes. Therefore, innovative charities like the Human Rights Foundation have started to offer donations via blockchain. Blockchain technology has the potential to establish profound trust and transparency in the donation process: Consumers can publicly track the progress of their donation at any time after deciding to donate. This ensures that the charity is not using donations against its original intent. Hence, the aim is to investigate the effect of blockchain-enabled transactions on the willingness to donate. Sample and Design: To investigate consumers' behavior, we use a scenario-based experiment. After removing participants (e.g., due to failed attention checks), 3192 potential donors participated (47.9% female, 62.4% bachelor or above). Procedure: We randomly assigned the participants to one of two scenarios. In all conditions, the participants read a scenario about a fictive charity organization called "Helper NPO." Afterward, the participants answered questions regarding their perception of the charity. Manipulation: The first scenario (n = 1405) represents a typical donation process, where consumers donate money without any option to track and trace. The second scenario (n = 1787) represents a donation process via blockchain, where consumers can track and trace their donations respectively. Using t-statistics, the findings demonstrate a positive effect of donating via blockchain on participants’ willingness to donate (mean difference = 0.667, p < .001, Cohen’s d effect size = 0.482). A mediation analysis shows significant effects for the mediation of transparency (Estimate = 0.199, p < .001), trust (Estimate = 0.144, p < .001), and transparency and trust (Estimate = 0.158, p < .001). The total effect of blockchain usage on participants’ willingness to donate (Estimate = 0.690, p < .001) consists of the direct effect (Estimate = 0.189, p < .001) and the indirect effects of transparency and trust (Estimate = 0.501, p < .001). Furthermore, consumers' affinity for technology moderates the direct effect of blockchain usage on participants' willingness to donate (Estimate = 0.150, p < .001). Donating via blockchain is a promising way for charities to engage consumers for several reasons: (1) Charities can emphasize trust and transparency in their advertising campaigns. (2) Established charities can target new customer segments by specifically engaging technology-affine consumers in the future. (3) Charities can raise international funds without previous barriers (e.g., setting up bank accounts). Nevertheless, increased transparency can also backfire (e.g., disclosure of costs). Such cases require further research.

Keywords: blockchain, social sector, transparency, trust

Procedia PDF Downloads 93
294 Distribution Patterns of the Renieramycin-M-Producing Blue Sponge, Xestospongia sp. (De Laubenfels, 1932) (Phylum: Porifera, Class: Demospongiae) in Puerto Galera, Oriental Mindoro, Philippines

Authors: Geminne Manzano, Clairecynth Yu, Lilibeth Salvador-Reyes, Viviene Santiago, Porfirio AliñO

Abstract:

The distribution and abundance patterns of many marine sessile organisms such as sponges vary among and within reefs. Determining the factors affecting its distribution is essential especially for organisms that produce secondary metabolites with pharmaceutical importance. In this study, the small-scale distribution patterns of the Philippine blue sponge, Xestospongia sp. in relation to some ecological factors were examined. The relationship between the renieramycin-M production and their benthic attributes were also determined. Ecological surveys were conducted on two stations with varying depth and exposure located in Oriental Mindoro, Philippines. Three 30 by 6m belt transect were used to assess the sponge abundance at each station. The substratum of the sponges was also characterized. Fish visual census observations were also taken together with the photo transect methods benthic surveys. Sponge samples were also collected for the extraction of Renieramycin-M and for further chemical analysis. Varying distribution patterns were observed to be attributed to the combination of different ecological and environmental factors. The amount of Renieramycin-production also varied in each station. The common substratum for blue sponges includes hard and soft corals, as well as, dead coral with algal patches. Blue sponges from exposed habitat frequently grow associated with massive and branching corals, Porites sp., while the most frequent substrate found on sheltered habitats is the coral Pavona sp. Exploring the influence of ecological and environmental parameters on the abundance and distribution of sponge assemblages provide ecological insights and their potential applications to pharmaceutical studies. The results of this study provide further impetus in pursuing studies into patterns and processes of the Philippine blue sponge, Xestospongia sp. distribution in relation to the chemical ecology of its secondary metabolites.

Keywords: distribution patterns, Porifera, Renieramycin-M, sponge assemblages, Xestospongia sp.

Procedia PDF Downloads 264
293 Modeling Driving Distraction Considering Psychological-Physical Constraints

Authors: Yixin Zhu, Lishengsa Yue, Jian Sun, Lanyue Tang

Abstract:

Modeling driving distraction in microscopic traffic simulation is crucial for enhancing simulation accuracy. Current driving distraction models are mainly derived from physical motion constraints under distracted states, in which distraction-related error terms are added to existing microscopic driver models. However, the model accuracy is not very satisfying, due to a lack of modeling the cognitive mechanism underlying the distraction. This study models driving distraction based on the Queueing Network Human Processor model (QN-MHP). This study utilizes the queuing structure of the model to perform task invocation and switching for distracted operation and control of the vehicle under driver distraction. Based on the assumption of the QN-MHP model about the cognitive sub-network, server F is a structural bottleneck. The latter information must wait for the previous information to leave server F before it can be processed in server F. Therefore, the waiting time for task switching needs to be calculated. Since the QN-MHP model has different information processing paths for auditory information and visual information, this study divides driving distraction into two types: auditory distraction and visual distraction. For visual distraction, both the visual distraction task and the driving task need to go through the visual perception sub-network, and the stimuli of the two are asynchronous, which is called stimulus on asynchrony (SOA), so when calculating the waiting time for switching tasks, it is necessary to consider it. In the case of auditory distraction, the auditory distraction task and the driving task do not need to compete for the server resources of the perceptual sub-network, and their stimuli can be synchronized without considering the time difference in receiving the stimuli. According to the Theory of Planned Behavior for drivers (TPB), this study uses risk entropy as the decision criterion for driver task switching. A logistic regression model is used with risk entropy as the independent variable to determine whether the driver performs a distraction task, to explain the relationship between perceived risk and distraction. Furthermore, to model a driver’s perception characteristics, a neurophysiological model of visual distraction tasks is incorporated into the QN-MHP, and executes the classical Intelligent Driver Model. The proposed driving distraction model integrates the psychological cognitive process of a driver with the physical motion characteristics, resulting in both high accuracy and interpretability. This paper uses 773 segments of distracted car-following in Shanghai Naturalistic Driving Study data (SH-NDS) to classify the patterns of distracted behavior on different road facilities and obtains three types of distraction patterns: numbness, delay, and aggressiveness. The model was calibrated and verified by simulation. The results indicate that the model can effectively simulate the distracted car-following behavior of different patterns on various roadway facilities, and its performance is better than the traditional IDM model with distraction-related error terms. The proposed model overcomes the limitations of physical-constraints-based models in replicating dangerous driving behaviors, and internal characteristics of an individual. Moreover, the model is demonstrated to effectively generate more dangerous distracted driving scenarios, which can be used to construct high-value automated driving test scenarios.

Keywords: computational cognitive model, driving distraction, microscopic traffic simulation, psychological-physical constraints

Procedia PDF Downloads 83
292 An Adaptive Decomposition for the Variability Analysis of Observation Time Series in Geophysics

Authors: Olivier Delage, Thierry Portafaix, Hassan Bencherif, Guillaume Guimbretiere

Abstract:

Most observation data sequences in geophysics can be interpreted as resulting from the interaction of several physical processes at several time and space scales. As a consequence, measurements time series in geophysics have often characteristics of non-linearity and non-stationarity and thereby exhibit strong fluctuations at all time-scales and require a time-frequency representation to analyze their variability. Empirical Mode Decomposition (EMD) is a relatively new technic as part of a more general signal processing method called the Hilbert-Huang transform. This analysis method turns out to be particularly suitable for non-linear and non-stationary signals and consists in decomposing a signal in an auto adaptive way into a sum of oscillating components named IMFs (Intrinsic Mode Functions), and thereby acts as a bank of bandpass filters. The advantages of the EMD technic are to be entirely data driven and to provide the principal variability modes of the dynamics represented by the original time series. However, the main limiting factor is the frequency resolution that may give rise to the mode mixing phenomenon where the spectral contents of some IMFs overlap each other. To overcome this problem, J. Gilles proposed an alternative entitled “Empirical Wavelet Transform” (EWT) which consists in building from the segmentation of the original signal Fourier spectrum, a bank of filters. The method used is based on the idea utilized in the construction of both Littlewood-Paley and Meyer’s wavelets. The heart of the method lies in the segmentation of the Fourier spectrum based on the local maxima detection in order to obtain a set of non-overlapping segments. Because linked to the Fourier spectrum, the frequency resolution provided by EWT is higher than that provided by EMD and therefore allows to overcome the mode-mixing problem. On the other hand, if the EWT technique is able to detect the frequencies involved in the original time series fluctuations, EWT does not allow to associate the detected frequencies to a specific mode of variability as in the EMD technic. Because EMD is closer to the observation of physical phenomena than EWT, we propose here a new technic called EAWD (Empirical Adaptive Wavelet Decomposition) based on the coupling of the EMD and EWT technics by using the IMFs density spectral content to optimize the segmentation of the Fourier spectrum required by EWT. In this study, EMD and EWT technics are described, then EAWD technic is presented. Comparison of results obtained respectively by EMD, EWT and EAWD technics on time series of ozone total columns recorded at Reunion island over [1978-2019] period is discussed. This study was carried out as part of the SOLSTYCE project dedicated to the characterization and modeling of the underlying dynamics of time series issued from complex systems in atmospheric sciences

Keywords: adaptive filtering, empirical mode decomposition, empirical wavelet transform, filter banks, mode-mixing, non-linear and non-stationary time series, wavelet

Procedia PDF Downloads 130
291 Mechanical and Tribological Performances of (Nb: H-D: a-C) Thin Films for Biomedical Applications

Authors: Sara Khamseh, Kambiz Javanruee, Hamid Khorsand

Abstract:

Plenty of metallic materials are used for biomedical applications like hip joints and screws. Besides, it is reported that metal platforms such as stainless steel show significant deterioration because of wear and friction. The surface of metal substrates has been coated with a variety of multicomponent coatings to prevail these problems. The carbon-based multicomponent coatings such as metal-added amorphous carbon and diamond coatings are crucially important because of their remarkable tribological performance and chemical stability. In the current study, H-D contained Nb: (a-C) multicomponent coatings (H-D: hexagonal diamond, a-C: amorphous carbon) coated on A 304 steel substrates using an unbalanced magnetron (UBM) sputtering system. The effects of Nb and H-D content and ID/IG ratio on microstructure, mechanical and tribological characteristics of (Nb: H-D: a-C) composite coatings were investigated. The results of Raman spectroscopy represented that a-C phase with a Graphite-like structure (GLC with high value of sp2 carbon bonding) is formed, and its domain size increased with increasing Nb content of the coatings. Moreover, the Nb played a catalyst for the formation of the H-D phase. The nanoindentation hardness value of the coatings ranged between ~17 to ~35 GPa and (Nb: H-D: a-C) composite coatings with more H-D content represented higher hardness and plasticity index. It seems that the existence of extra-hard H-D particles straightly increased hardness. The tribological performance of the coatings was evaluated using the pin-on-disc method under the wet environment of SBF (Simulated Body Fluid). The COF value of the (Nb: H-D: a-C) coatings decreased with an increasing ID/IG ratio. The lower coefficient of friction is a result of the lamelliform array of graphitic domains. Also, the wear rate of the coatings decreased with increasing H-D content of the coatings. Based on the literature, a-C coatings with high hardness and H3/E2 ratio represent lower wear rates and better tribological performance. According to the nanoindentation analysis, hardness and H3/E2 ratio of (Nb: H-D: a-C) multicomponent coatings increased with increasing H-D content, which in turn decreased the wear rate of the coatings. The mechanical and tribological potency of (Nb: H-D: a-C) composite coatings on A 304 steel substrates paved the way for the development of innovative advanced coatings to ameliorate the performance of A 304 steel for biomedical applications.

Keywords: COF, mechanical properties, (Nb: H-D: a-C) coatings, wear rate

Procedia PDF Downloads 95
290 Examining Experiences of QTBIPOC Disabled Students in Canadian Post-Secondary Institutions

Authors: Manchari Paranthahan

Abstract:

Higher education has often presented barriers to many communities as a result of its colonial roots. While higher education was initially created for white cis-males, student populations have become more diverse in the past few decades. Despite this increase in diversity, barriers like rising costs and hostile education settings continue to make higher education hard to access for certain demographics. These barriers and limitations are compounded for students who are intersectionality marginalized, such as Queer and Trans Black, Indigenous and People of Colour (QTBIPOC) Disabled students. As of 2021-2022, only 57.5% of the Canadian population between the ages of 25 - 64 held a college or university credential, with only 32.9% holding a bachelor’s degree or higher. In that same time frame, only 0.64% of the students who successfully completed a higher education program identified as transgender or nonbinary. QTBIPOC Disabled students experience diverse forms of oppression while navigating education systems, often preventing them from completing their education successfully. This research project will investigate the complex experiences of intersectional marginalization of QTBIPOC Disabled students in Canadian post-secondary education systems. Through this investigation, this research seeks to reimagine more inclusive and accessible education systems in Canada and beyond. The social and academic experiences of QTBIPOC Disabled students in education systems are largely absent from scholarly literature, speaking to their continued marginalization and erasure from academic discourses. The lack of representation for this community in academia reinforces the idea that there is no space for marginalized bodies in further education, a discriminatory belief that this research project aims to investigate and reframe with this project. This research study will be informed by Critical Race theory, Queer Theory and Critical Disability Theories. Through a blend of critical narrative ethnography and ethnodrama for my methodological framing. Using these methodologies will speak to the intersecting factors that impact the experiences that QTBIPOC Disabled students have in education systems while offering space to analyze and create new systems of learning that benefits all students.

Keywords: QTBIPOC, queer, disability, pedagogy

Procedia PDF Downloads 15
289 Administrators' Information Management Capacity and Decision-Making Effectiveness on Staff Promotion in the Teaching Service Commissions in South – West, Nigeria

Authors: Olatunji Sabitu Alimi

Abstract:

This study investigated the extent to which administrators’ information storage, retrieval and processing capacities influence decisions on staff promotion in the Teaching Service Commissions (TESCOMs) in The South-West, Nigeria. One research question and two research hypotheses were formulated and tested respectively at 0.05 level of significance. The study used the descriptive research of the survey type. One hundred (100) staff on salary grade level 09 constituted the sample. Multi- stage, stratified and simple random sampling techniques were used to select 100 staff from the TESCOMs in The South-West, Nigeria. Two questionnaires titled Administrators’ Information Storage, Retrieval and Processing Capacities (AISRPC), and Staff Promotion Effectiveness (SPE) were used for data collection. The inventory was validated and subjected to test-re-test and reliability coefficient of r = 0.79 was obtained. The data were collected and analyzed using Pearson Product Moment Correlation coefficient and simple percentage. The study found that Administrators at TESCOM stored their information in files, hard copies, soft copies, open registry and departmentally in varying degrees while they also processed information manually and through electronics for decision making. In addition, there is a significant relationship between administrators’ information storage and retrieval capacities in the TESCOMs in South – West, Nigeria, (r cal = 0.598 > r table = 0.195). Furthermore, administrators’ information processing capacity and staff promotion effectiveness were found to be significantly related (r cal = 0.209 > r table = 0.195 at 0.05 level of significance). The study recommended that training, seminars, workshops should be organized for administrators on information management, while educational organizations should provide Information Management Technology (ICT) equipment for the administrators in the TESCOMs. The staff of TESCOM should be promoted having satisfied the promotion criteria such as spending required number of years on a grade level, a clean record of service and vacancy.

Keywords: information processing capacity, staff promotion effectiveness, teaching service commission, Nigeria

Procedia PDF Downloads 530
288 A Comprehensive Survey on Machine Learning Techniques and User Authentication Approaches for Credit Card Fraud Detection

Authors: Niloofar Yousefi, Marie Alaghband, Ivan Garibay

Abstract:

With the increase of credit card usage, the volume of credit card misuse also has significantly increased, which may cause appreciable financial losses for both credit card holders and financial organizations issuing credit cards. As a result, financial organizations are working hard on developing and deploying credit card fraud detection methods, in order to adapt to ever-evolving, increasingly sophisticated defrauding strategies and identifying illicit transactions as quickly as possible to protect themselves and their customers. Compounding on the complex nature of such adverse strategies, credit card fraudulent activities are rare events compared to the number of legitimate transactions. Hence, the challenge to develop fraud detection that are accurate and efficient is substantially intensified and, as a consequence, credit card fraud detection has lately become a very active area of research. In this work, we provide a survey of current techniques most relevant to the problem of credit card fraud detection. We carry out our survey in two main parts. In the first part, we focus on studies utilizing classical machine learning models, which mostly employ traditional transnational features to make fraud predictions. These models typically rely on some static physical characteristics, such as what the user knows (knowledge-based method), or what he/she has access to (object-based method). In the second part of our survey, we review more advanced techniques of user authentication, which use behavioral biometrics to identify an individual based on his/her unique behavior while he/she is interacting with his/her electronic devices. These approaches rely on how people behave (instead of what they do), which cannot be easily forged. By providing an overview of current approaches and the results reported in the literature, this survey aims to drive the future research agenda for the community in order to develop more accurate, reliable and scalable models of credit card fraud detection.

Keywords: Credit Card Fraud Detection, User Authentication, Behavioral Biometrics, Machine Learning, Literature Survey

Procedia PDF Downloads 111
287 Career Guidance System Using Machine Learning

Authors: Mane Darbinyan, Lusine Hayrapetyan, Elen Matevosyan

Abstract:

Artificial Intelligence in Education (AIED) has been created to help students get ready for the workforce, and over the past 25 years, it has grown significantly, offering a variety of technologies to support academic, institutional, and administrative services. However, this is still challenging, especially considering the labor market's rapid change. While choosing a career, people face various obstacles because they do not take into consideration their own preferences, which might lead to many other problems like shifting jobs, work stress, occupational infirmity, reduced productivity, and manual error. Besides preferences, people should properly evaluate their technical and non-technical skills, as well as their personalities. Professional counseling has become a difficult undertaking for counselors due to the wide range of career choices brought on by changing technological trends. It is necessary to close this gap by utilizing technology that makes sophisticated predictions about a person's career goals based on their personality. Hence, there is a need to create an automated model that would help in decision-making based on user inputs. Improving career guidance can be achieved by embedding machine learning into the career consulting ecosystem. There are various systems of career guidance that work based on the same logic, such as the classification of applicants, matching applications with appropriate departments or jobs, making predictions, and providing suitable recommendations. Methodologies like KNN, Neural Networks, K-means clustering, D-Tree, and many other advanced algorithms are applied in the fields of data and compute some data, which is helpful to predict the right careers. Besides helping users with their career choice, these systems provide numerous opportunities which are very useful while making this hard decision. They help the candidate to recognize where he/she specifically lacks sufficient skills so that the candidate can improve those skills. They are also capable to offer an e-learning platform, taking into account the user's lack of knowledge. Furthermore, users can be provided with details on a particular job, such as the abilities required to excel in that industry.

Keywords: career guidance system, machine learning, career prediction, predictive decision, data mining, technical and non-technical skills

Procedia PDF Downloads 76
286 Career Guidance System Using Machine Learning

Authors: Mane Darbinyan, Lusine Hayrapetyan, Elen Matevosyan

Abstract:

Artificial Intelligence in Education (AIED) has been created to help students get ready for the workforce, and over the past 25 years, it has grown significantly, offering a variety of technologies to support academic, institutional, and administrative services. However, this is still challenging, especially considering the labor market's rapid change. While choosing a career, people face various obstacles because they do not take into consideration their own preferences, which might lead to many other problems like shifting jobs, work stress, occupational infirmity, reduced productivity, and manual error. Besides preferences, people should evaluate properly their technical and non-technical skills, as well as their personalities. Professional counseling has become a difficult undertaking for counselors due to the wide range of career choices brought on by changing technological trends. It is necessary to close this gap by utilizing technology that makes sophisticated predictions about a person's career goals based on their personality. Hence, there is a need to create an automated model that would help in decision-making based on user inputs. Improving career guidance can be achieved by embedding machine learning into the career consulting ecosystem. There are various systems of career guidance that work based on the same logic, such as the classification of applicants, matching applications with appropriate departments or jobs, making predictions, and providing suitable recommendations. Methodologies like KNN, neural networks, K-means clustering, D-Tree, and many other advanced algorithms are applied in the fields of data and compute some data, which is helpful to predict the right careers. Besides helping users with their career choice, these systems provide numerous opportunities which are very useful while making this hard decision. They help the candidate to recognize where he/she specifically lacks sufficient skills so that the candidate can improve those skills. They are also capable of offering an e-learning platform, taking into account the user's lack of knowledge. Furthermore, users can be provided with details on a particular job, such as the abilities required to excel in that industry.

Keywords: career guidance system, machine learning, career prediction, predictive decision, data mining, technical and non-technical skills

Procedia PDF Downloads 65
285 Identifying Necessary Words for Understanding Academic Articles in English as a Second or a Foreign Language

Authors: Stephen Wagman

Abstract:

This paper identifies three common structures in English sentences that are important for understanding academic texts, regardless of the characteristics or background of the readers or whether they are reading English as a second or a foreign language. Adapting a model from the Humanities, the explication of texts used in literary studies, the paper analyses sample sentences to reveal structures that enable the reader not only to decide which words are necessary for understanding the main ideas but to make the decision without knowing the meaning of the words. By their very syntax noun structures point to the key word for understanding them. As a rule, the key noun is followed by easily identifiable prepositions, relative pronouns, or verbs and preceded by single adjectives. With few exceptions, the modifiers are unnecessary for understanding the idea of the sentence. In addition, sentences are often structured by lists in which the items frequently consist of parallel groups of words. The principle of a list is that all the items are similar in meaning and it is not necessary to understand all of the items to understand the point of the list. This principle is especially important when the items are long or there is more than one list in the same sentence. The similarity in meaning of these items enables readers to reduce sentences that are hard to grasp to an understandable core without excessive use of a dictionary. Finally, the idea of subordination and the identification of the subordinate parts of sentences through connecting words makes it possible for readers to focus on main ideas without having to sift through the less important and more numerous secondary structures. Sometimes a main idea requires a subordinate one to complete its meaning, but usually, subordinate ideas are unnecessary for understanding the main point of the sentence and its part in the development of the argument from sentence to sentence. Moreover, the connecting words themselves indicate the functions of the subordinate structures. These most frequently show similarity and difference or reasons and results. Recognition of all of these structures can not only enable students to read more efficiently but to focus their attention on the development of the argument and this rather than a multitude of unknown vocabulary items, the repetition in lists, or the subordination in sentences are the one necessary element for comprehension of academic articles.

Keywords: development of the argument, lists, noun structures, subordination

Procedia PDF Downloads 242