Search results for: data reconstruction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24540

Search results for: data reconstruction

24270 Airport Pavement Crack Measurement Systems and Crack Density for Pavement Evaluation

Authors: Ali Ashtiani, Hamid Shirazi

Abstract:

This paper reviews the status of existing practice and research related to measuring pavement cracking and using crack density as a pavement surface evaluation protocol. Crack density for pavement evaluation is currently not widely used within the airport community and its use by the highway community is limited. However, surface cracking is a distress that is closely monitored by airport staff and significantly influences the development of maintenance, rehabilitation and reconstruction plans for airport pavements. Therefore crack density has the potential to become an important indicator of pavement condition if the type, severity and extent of surface cracking can be accurately measured. A pavement distress survey is an essential component of any pavement assessment. Manual crack surveying has been widely used for decades to measure pavement performance. However, the accuracy and precision of manual surveys can vary depending upon the surveyor and performing surveys may disrupt normal operations. Given the variability of manual surveys, this method has shown inconsistencies in distress classification and measurement. This can potentially impact the planning for pavement maintenance, rehabilitation and reconstruction and the associated funding strategies. A substantial effort has been devoted for the past 20 years to reduce the human intervention and the error associated with it by moving toward automated distress collection methods. The automated methods refer to the systems that identify, classify and quantify pavement distresses through processes that require no or very minimal human intervention. This principally involves the use of a digital recognition software to analyze and characterize pavement distresses. The lack of established protocols for measurement and classification of pavement cracks captured using digital images is a challenge to developing a reliable automated system for distress assessment. Variations in types and severity of distresses, different pavement surface textures and colors and presence of pavement joints and edges all complicate automated image processing and crack measurement and classification. This paper summarizes the commercially available systems and technologies for automated pavement distress evaluation. A comprehensive automated pavement distress survey involves collection, interpretation, and processing of the surface images to identify the type, quantity and severity of the surface distresses. The outputs can be used to quantitatively calculate the crack density. The systems for automated distress survey using digital images reviewed in this paper can assist the airport industry in the development of a pavement evaluation protocol based on crack density. Analysis of automated distress survey data can lead to a crack density index. This index can be used as a means of assessing pavement condition and to predict pavement performance. This can be used by airport owners to determine the type of pavement maintenance and rehabilitation in a more consistent way.

Keywords: airport pavement management, crack density, pavement evaluation, pavement management

Procedia PDF Downloads 164
24269 Government Big Data Ecosystem: A Systematic Literature Review

Authors: Syed Iftikhar Hussain Shah, Vasilis Peristeras, Ioannis Magnisalis

Abstract:

Data that is high in volume, velocity, veracity and comes from a variety of sources is usually generated in all sectors including the government sector. Globally public administrations are pursuing (big) data as new technology and trying to adopt a data-centric architecture for hosting and sharing data. Properly executed, big data and data analytics in the government (big) data ecosystem can be led to data-driven government and have a direct impact on the way policymakers work and citizens interact with governments. In this research paper, we conduct a systematic literature review. The main aims of this paper are to highlight essential aspects of the government (big) data ecosystem and to explore the most critical socio-technical factors that contribute to the successful implementation of government (big) data ecosystem. The essential aspects of government (big) data ecosystem include definition, data types, data lifecycle models, and actors and their roles. We also discuss the potential impact of (big) data in public administration and gaps in the government data ecosystems literature. As this is a new topic, we did not find specific articles on government (big) data ecosystem and therefore focused our research on various relevant areas like humanitarian data, open government data, scientific research data, industry data, etc.

Keywords: applications of big data, big data, big data types. big data ecosystem, critical success factors, data-driven government, egovernment, gaps in data ecosystems, government (big) data, literature review, public administration, systematic review

Procedia PDF Downloads 180
24268 A Machine Learning Decision Support Framework for Industrial Engineering Purposes

Authors: Anli Du Preez, James Bekker

Abstract:

Data is currently one of the most critical and influential emerging technologies. However, the true potential of data is yet to be exploited since, currently, about 1% of generated data are ever actually analyzed for value creation. There is a data gap where data is not explored due to the lack of data analytics infrastructure and the required data analytics skills. This study developed a decision support framework for data analytics by following Jabareen’s framework development methodology. The study focused on machine learning algorithms, which is a subset of data analytics. The developed framework is designed to assist data analysts with little experience, in choosing the appropriate machine learning algorithm given the purpose of their application.

Keywords: Data analytics, Industrial engineering, Machine learning, Value creation

Procedia PDF Downloads 137
24267 Providing Security to Private Cloud Using Advanced Encryption Standard Algorithm

Authors: Annapureddy Srikant Reddy, Atthanti Mahendra, Samala Chinni Krishna, N. Neelima

Abstract:

In our present world, we are generating a lot of data and we, need a specific device to store all these data. Generally, we store data in pen drives, hard drives, etc. Sometimes we may loss the data due to the corruption of devices. To overcome all these issues, we implemented a cloud space for storing the data, and it provides more security to the data. We can access the data with just using the internet from anywhere in the world. We implemented all these with the java using Net beans IDE. Once user uploads the data, he does not have any rights to change the data. Users uploaded files are stored in the cloud with the file name as system time and the directory will be created with some random words. Cloud accepts the data only if the size of the file is less than 2MB.

Keywords: cloud space, AES, FTP, NetBeans IDE

Procedia PDF Downloads 174
24266 Potential of Mineral Composition Reconstruction for Monitoring the Performance of an Iron Ore Concentration Plant

Authors: Maryam Sadeghi, Claude Bazin, Daniel Hodouin, Laura Perez Barnuevo

Abstract:

The performance of a separation process is usually evaluated using performance indices calculated from elemental assays readily available from the chemical analysis laboratory. However, the separation process performance is essentially related to the properties of the minerals that carry the elements and not those of the elements. Since elements or metals can be carried by valuable and gangue minerals in the ore and that each mineral responds differently to a mineral processing method, the use of only elemental assays could lead to erroneous or uncertain conclusions on the process performance. This paper discusses the advantages of using performance indices calculated from minerals content, such as minerals recovery, for process performance assessments. A method is presented that uses elemental assays to estimate the minerals content of the solids in various process streams. The method combines the stoichiometric composition of the minerals and constraints of mass conservation for the minerals through the concentration process to estimate the minerals content from elemental assays. The advantage of assessing a concentration process using mineral based performance indices is illustrated for an iron ore concentration circuit.

Keywords: data reconciliation, iron ore concentration, mineral composition, process performance assessment

Procedia PDF Downloads 177
24265 An Optimal Control Method for Reconstruction of Topography in Dam-Break Flows

Authors: Alia Alghosoun, Nabil El Moçayd, Mohammed Seaid

Abstract:

Modeling dam-break flows over non-flat beds requires an accurate representation of the topography which is the main source of uncertainty in the model. Therefore, developing robust and accurate techniques for reconstructing topography in this class of problems would reduce the uncertainty in the flow system. In many hydraulic applications, experimental techniques have been widely used to measure the bed topography. In practice, experimental work in hydraulics may be very demanding in both time and cost. Meanwhile, computational hydraulics have served as an alternative for laboratory and field experiments. Unlike the forward problem, the inverse problem is used to identify the bed parameters from the given experimental data. In this case, the shallow water equations used for modeling the hydraulics need to be rearranged in a way that the model parameters can be evaluated from measured data. However, this approach is not always possible and it suffers from stability restrictions. In the present work, we propose an adaptive optimal control technique to numerically identify the underlying bed topography from a given set of free-surface observation data. In this approach, a minimization function is defined to iteratively determine the model parameters. The proposed technique can be interpreted as a fractional-stage scheme. In the first stage, the forward problem is solved to determine the measurable parameters from known data. In the second stage, the adaptive control Ensemble Kalman Filter is implemented to combine the optimality of observation data in order to obtain the accurate estimation of the topography. The main features of this method are on one hand, the ability to solve for different complex geometries with no need for any rearrangements in the original model to rewrite it in an explicit form. On the other hand, its achievement of strong stability for simulations of flows in different regimes containing shocks or discontinuities over any geometry. Numerical results are presented for a dam-break flow problem over non-flat bed using different solvers for the shallow water equations. The robustness of the proposed method is investigated using different numbers of loops, sensitivity parameters, initial samples and location of observations. The obtained results demonstrate high reliability and accuracy of the proposed techniques.

Keywords: erodible beds, finite element method, finite volume method, nonlinear elasticity, shallow water equations, stresses in soil

Procedia PDF Downloads 106
24264 Antagonist Coactivation in Athletes Following Anterior Cruciate Ligament Reconstruction

Authors: Milad Pirali, Sohrab Keyhani, Mohhamad Ali Sanjari, Ali Ashraf Jamshidi

Abstract:

Purpose: The effect of hamstring antagonist activity on the knee extensors torque of the Anterior Cruciate Ligament reconstruction (ACLR) is not clear and persistent muscle weakness is common after ACLR. Hamstring activation when acting as antagonist is considered very important for knee strengths. Therefore the purpose of this study was to examine hamstring antagonist coactivation during maximal effort of the isokinetic knee extension in ACLR athletes with hamstring autograft. Materials and Methods: We enrolled 20 professional athletes who underwent primary ACLR (hamstring tendon autograft)with 6-24 months postoperative and 20 healthy subjects as control group. Each subjects performed maximal effort isokinetic knee extension and flexion in 60/˚ s and 180/˚ s velocities for the involved and uninvolved limb. Synchronously, surface electromyography (EMG) was recorded of vastus medialis (VM), vastus lateralis (VL), rectus femoris (RF) and biceps femoris (BF). The antagonist integrated EMG (IEMG) values were normalized to the IEMG of the same muscle during maximal isokinetic eccentric effort at the same velocities and ROM. Results: A one-way analysis of variance designs shows significantly greater IEMG coactivation of hamstring and decreased activation of Vm in ACLR when compared to uninvolved and control group leg in 60/˚ s and 180/˚ s velocities. Likewise peak torque to body weight was decreased in ACLR compared to uninvolved and control group during knee extension in both velocities (p < 0.05). Conclusions: Decreased extensors moment caused by decreased quadriceps inhibition and increased hamstring coactivation. In addition, these result indicated to decrease of motor unit recruitment in the VM (as a kinesiologicmonitore of the knee). It is appearing that strengthening of the quadriceps to be an important for rehabilitation program after ACLR for preparation in athletes endeavors. Therefore, we suggest that having more emphasis and focus on quadriceps strength and less emphasis on hamstring following ACLR.

Keywords: ACLR-coactivation, dynamometry, electromyography, isokinetic

Procedia PDF Downloads 224
24263 Business Intelligence for Profiling of Telecommunication Customer

Authors: Rokhmatul Insani, Hira Laksmiwati Soemitro

Abstract:

Business Intelligence is a methodology that exploits the data to produce information and knowledge systematically, business intelligence can support the decision-making process. Some methods in business intelligence are data warehouse and data mining. A data warehouse can store historical data from transactional data. For data modelling in data warehouse, we apply dimensional modelling by Kimball. While data mining is used to extracting patterns from the data and get insight from the data. Data mining has many techniques, one of which is segmentation. For profiling of telecommunication customer, we use customer segmentation according to customer’s usage of services, customer invoice and customer payment. Customers can be grouped according to their characteristics and can be identified the profitable customers. We apply K-Means Clustering Algorithm for segmentation. The input variable for that algorithm we use RFM (Recency, Frequency and Monetary) model. All process in data mining, we use tools IBM SPSS modeller.

Keywords: business intelligence, customer segmentation, data warehouse, data mining

Procedia PDF Downloads 443
24262 Reconstruction of Binary Matrices Satisfying Neighborhood Constraints by Simulated Annealing

Authors: Divyesh Patel, Tanuja Srivastava

Abstract:

This paper considers the NP-hard problem of reconstructing binary matrices satisfying exactly-1-4-adjacency constraint from its row and column projections. This problem is formulated into a maximization problem. The objective function gives a measure of adjacency constraint for the binary matrices. The maximization problem is solved by the simulated annealing algorithm and experimental results are presented.

Keywords: discrete tomography, exactly-1-4-adjacency, simulated annealing, binary matrices

Procedia PDF Downloads 374
24261 Privatising Higher Education: Imparting Quality in Academics

Authors: Manish Khanna

Abstract:

Higher education seeks to preserve, transmit and advance knowledge. It is one of the most important instruments of change and progress. The observation of Kothari Commission (1964-66) is true even today; The destiny of India is now being shaped in her classrooms. This, we believe, is no more rhetoric. In the world based on science and technology it is education that determines the level of prosperity, welfare, and security of the people. On the quality and number of persons coming out of our schools and colleges will depend our success in the great enterprise of national reconstruction.

Keywords: higher education, quality in academics, Kothari commission, privatising higher education

Procedia PDF Downloads 443
24260 Imputation Technique for Feature Selection in Microarray Data Set

Authors: Younies Saeed Hassan Mahmoud, Mai Mabrouk, Elsayed Sallam

Abstract:

Analysing DNA microarray data sets is a great challenge, which faces the bioinformaticians due to the complication of using statistical and machine learning techniques. The challenge will be doubled if the microarray data sets contain missing data, which happens regularly because these techniques cannot deal with missing data. One of the most important data analysis process on the microarray data set is feature selection. This process finds the most important genes that affect certain disease. In this paper, we introduce a technique for imputing the missing data in microarray data sets while performing feature selection.

Keywords: DNA microarray, feature selection, missing data, bioinformatics

Procedia PDF Downloads 533
24259 PDDA: Priority-Based, Dynamic Data Aggregation Approach for Sensor-Based Big Data Framework

Authors: Lutful Karim, Mohammed S. Al-kahtani

Abstract:

Sensors are being used in various applications such as agriculture, health monitoring, air and water pollution monitoring, traffic monitoring and control and hence, play the vital role in the growth of big data. However, sensors collect redundant data. Thus, aggregating and filtering sensors data are significantly important to design an efficient big data framework. Current researches do not focus on aggregating and filtering data at multiple layers of sensor-based big data framework. Thus, this paper introduces (i) three layers data aggregation and framework for big data and (ii) a priority-based, dynamic data aggregation scheme (PDDA) for the lowest layer at sensors. Simulation results show that the PDDA outperforms existing tree and cluster-based data aggregation scheme in terms of overall network energy consumptions and end-to-end data transmission delay.

Keywords: big data, clustering, tree topology, data aggregation, sensor networks

Procedia PDF Downloads 302
24258 Subjective Well-Being through Coaching Process

Authors: Pendar Fazel

Abstract:

Well-being is a good or satisfactory condition of existence; a state characterized by health, happiness, and prosperity. Well-being of people is correlated with, the cognitive, social, emotional, and physical aspect of their personality. Subjective well-being, people’s emotional and cognitive evaluations of their lives, includes what lay people call happiness, peace, fulfillment, and life satisfaction. Unfortunately in this period of time people are under the pressure of financial, social problems, and other stress factors which made them vulnerable, and their well-being is threatened. Personal Coaching as a holistic orientation and novel approach is ideal for the present century which help people, to find balance, enjoyment and meaning in their lives as well as improving performance, skills and effectiveness. The aim of the present article besides introducing the personal coaching is determining how personal coaching can positively effects on subjective well-being, under this aim we tend to describe how coaching impact on the cognitive and emotional reconstruction. Present qualitative research is descriptive analytic study, which data gathered by manual library research and search within authentic article through internet; analyzed personal coaching which integrated different views into an operational one helps people promote self-awareness as well as evaluate, emotional and cognitive aspect of their personality and provide appropriate subjective well-being.

Keywords: subjective well-being, coaching, well-being, positive psychology, personal growth

Procedia PDF Downloads 503
24257 3D Microscopy, Image Processing, and Analysis of Lymphangiogenesis in Biological Models

Authors: Thomas Louis, Irina Primac, Florent Morfoisse, Tania Durre, Silvia Blacher, Agnes Noel

Abstract:

In vitro and in vivo lymphangiogenesis assays are essential for the identification of potential lymphangiogenic agents and the screening of pharmacological inhibitors. In the present study, we analyse three biological models: in vitro lymphatic endothelial cell spheroids, in vivo ear sponge assay, and in vivo lymph node colonisation by tumour cells. These assays provide suitable 3D models to test pro- and anti-lymphangiogenic factors or drugs. 3D images were acquired by confocal laser scanning and light sheet fluorescence microscopy. Virtual scan microscopy followed by 3D reconstruction by image aligning methods was also used to obtain 3D images of whole large sponge and ganglion samples. 3D reconstruction, image segmentation, skeletonisation, and other image processing algorithms are described. Fixed and time-lapse imaging techniques are used to analyse lymphatic endothelial cell spheroids behaviour. The study of cell spatial distribution in spheroid models enables to detect interactions between cells and to identify invasion hierarchy and guidance patterns. Global measurements such as volume, length, and density of lymphatic vessels are measured in both in vivo models. Branching density and tortuosity evaluation are also proposed to determine structure complexity. Those properties combined with vessel spatial distribution are evaluated in order to determine lymphangiogenesis extent. Lymphatic endothelial cell invasion and lymphangiogenesis were evaluated under various experimental conditions. The comparison of these conditions enables to identify lymphangiogenic agents and to better comprehend their roles in the lymphangiogenesis process. The proposed methodology is validated by its application on the three presented models.

Keywords: 3D image segmentation, 3D image skeletonisation, cell invasion, confocal microscopy, ear sponges, light sheet microscopy, lymph nodes, lymphangiogenesis, spheroids

Procedia PDF Downloads 340
24256 An Interpretative Phenomenological Analysis on the Concept of Friends of Children in Conflict with the Law

Authors: Karla Kristine Bay, Jovie Ann Gabin, Allana Joyce Sasotona

Abstract:

This research employed an Interpretative Phenomenological Analysis to explore the experiences of Children in Conflict with the Law (CICL) which gave light to their concept of ‘friends’. Derived from this context are the following objectives of the study: 1) determining the differentiation of the forms of friends of the CICL; 2) presenting the process of attachment towards detachment in the formation of friendship; and 3) discussing the experiences, and reflections of the CICL on the ‘self’ out of their encounter with friendship. Using the data gathered from the individual drawings of the CICL of their representations of the self, family, friends, community, and Bahay Kalinga as subjects in the meaning-making process utilizing Filipino Psychology methods of pagtatanong-tanong (interview), and pakikipagkwentuhan (conversation), data analysis produced a synthesis of seventeen individual cases. Overall results generated three superordinate themes on the differentiation of the forms of friends which include friends with good influences, friends with bad influences, and friends within the family. While two superordinate themes were produced on the process of attachment towards detachment, namely social, emotional, and psychological experiences on the process of attachment, and emotional and psychological experiences on the process of detachment. Lastly, two superordinate themes were created on the experiences, and reflections of the CICL on the ‘self’ out of their encounter with friendship. This consists of the recognition of the ‘self’ as a responsible agent in developing healthy relationships between the self and others, and reconstruction of the self from the collective experiences of healing, forgiveness, and acceptance. These findings, together with supporting theories discussed the impact of friendship on the emergence of criminal behavior and other dispositions; springing from the child’s dissociation from the family that led to finding belongingness from an external group called friends.

Keywords: children in conflict with the law, criminal behavior, friends, interpretative phenomenological analysis

Procedia PDF Downloads 213
24255 Laboratory Investigation of the Pavement Condition in Lebanon: Implementation of Reclaimed Asphalt Pavement in the Base Course and Asphalt Layer

Authors: Marinelle El-Khoury, Lina Bouhaya, Nivine Abbas, Hassan Sleiman

Abstract:

The road network in the north of Lebanon is a prime example of the lack of pavement design and execution in Lebanon.  These roads show major distresses and hence, should be tested and evaluated. The aim of this research is to investigate and determine the deficiencies in road surface design in Lebanon, and to propose an environmentally friendly asphalt mix design. This paper consists of several parts: (i) evaluating pavement performance and structural behavior, (ii) identifying the distresses using visual examination followed by laboratory tests, (iii) deciding the optimal solution where rehabilitation or reconstruction is required and finally, (iv) identifying a sustainable method, which uses recycled material in the proposed mix. The asphalt formula contains Reclaimed Asphalt Pavement (RAP) in the base course layer and in the asphalt layer. Visual inspection of the roads in Tripoli shows that these roads face a high level of distress severity. Consequently, the pavement should be reconstructed rather than simply rehabilitated. Coring was done to determine the pavement layer thickness. The results were compared to the American Association of State Highway and Transportation Officials (AASHTO) design methodology and showed that the existing asphalt thickness is lower than the required asphalt thickness. Prior to the pavement reconstruction, the road materials were tested according to the American Society for Testing and Materials (ASTM) specification to identify whether the materials are suitable. Accordingly, the ASTM tests that were performed on the base course are Sieve analysis, Atterberg limits, modified proctor, Los Angeles, and California Bearing Ratio (CBR) tests. Results show a CBR value higher than 70%. Hence, these aggregates could be used as a base course layer. The asphalt layer was also tested and the results of the Marshall flow and stability tests meet the ASTM specifications. In the last section, an environmentally friendly mix was proposed. An optimal RAP percentage of 30%, which produced a well graded base course and asphalt mix, was determined through a series of trials.

Keywords: asphalt mix, reclaimed asphalt pavement, California bearing ratio, sustainability

Procedia PDF Downloads 95
24254 The Global Economic System and the Third World Development

Authors: Monday Dickson

Abstract:

Shortly before the end of the second world war, allied leaders and other western powers designed an economic regime that would foster, among other things, global economic reconstruction, prosperity and overall development of countries of the world. They founded both the World Bank and the International Monetary Fund (IMF), with a general consensus that while the latter should specialize in monitoring global and national economies and acting as a lender of last resort, the former should focus on fighting poverty and promoting development. In setting the rules for world trade, the General Agreement on Trade and Tariffs (GATT) evolved into the World Trade Organisation (WTO). This paper, therefore, examines the impact of the activities of these institutions on the transformation and development aspirations of countries of the Third World. The study adopts the descriptive and analytical methods of investigation and derived relevant secondary data from books, journal articles, encyclopedia as well as reports from countries of the Third World. Findings show that rather than fostering poverty reduction and overall development as envisaged, the activities of global economy system leads to the “development of underdevelopment” of the Third World Countries. The strategic options that are available to countries of the Third World derived from the ability of the national governments to develop programmes of systematic exploration and exploitation of vital indices of relations with strategic countries to advance their development agenda.

Keywords: development, global economic system, prosperity, third world

Procedia PDF Downloads 407
24253 Reconstruction and Renewal of Traditional Houses and its Impact on Tourism Development in Rasht

Authors: Parvaneh Ziviyar, Simin Armaghan

Abstract:

Traditional house in Rasht contains monuments and heritage of ancestors who once lived in these houses. These houses represent the customs, culture and lifestyle of the people of Rasht and bridge the gap between modern people and their past that is being forgotten. Maintenance of the buildings and architectural heritage together with their unique architecture and climatic related construction has an important role in tourism attraction and sustainable development. The purpose of this study was to develop a new definition of vacation shacks that is different with the definition of Cultural Heritage Organization. The place to stay and visit that is rebuilt or renovated based on traditional architectural style of Rasht and yet provides modern amenities so that it would not undermine indigenous traditional sense of the house. Data collection for this study is based on review of literature and field study. Results and the statistics of this study will prove that the research hypothesis is supported and there is a correlation between traditional houses of Rasht, as tourism–accommodation place and tourist attraction. It also indicates the capability and potential of these ancient monuments in the introduction of the culture of this land, and calling people and many tourists come to visit and stay in such places.

Keywords: architecture, traditional houses, vacation shacks, tourism

Procedia PDF Downloads 251
24252 Effect of Retained Posterior Horn of Medial Meniscus on Functional Outcome of ACL Reconstructed Knees

Authors: Kevin Syam, Devendra K. Chauhan, Mandeep Singh Dhillon

Abstract:

Background: The posterior horn of medial meniscus (PHMM) is a secondary stabilizer against anterior translation of tibia. Cadaveric studies have revealed increased strain on the ACL graft and greater instrumented laxity in Posterior horn deficient knees. Clinical studies have shown higher prevalence of radiological OA after ACL reconstruction combined with menisectomy. However, functional outcomes in ACL reconstructed knee in the absence of Posterior horn is less discussed, and specific role of posterior horn is ill-documented. This study evaluated functional and radiological outcomes in posterior horn preserved and posterior horn sacrificed ACL reconstructed knees. Materials: Of the 457 patients who had ACL reconstruction done over a 6 year period, 77 cases with minimum follow up of 18 months were included in the study after strict exclusion criteria (associated lateral meniscus injury, other ligamentous injuries, significant cartilage degeneration, repeat injury and contralateral knee injuries were excluded). 41 patients with intact menisci were compared with 36 patients with absent posterior horn of medial meniscus. Radiological and clinical tests for instability were conducted, and knees were evaluated using subjective International Knee Documentation Committee (IKDC) score and the Orthopadische Arbeitsgruppe Knie score (OAK). Results: We found a trend towards significantly better overall outcome (OAK) in cases with intact PHMM at average follow-up of 43.03 months (p value 0.082). Cases with intact PHMM had significantly better objective stability (p value 0.004). No significant differences were noted in the subjective IKDC score (p value 0.526) and the functional OAK outcome (category D) (p value 0.363). More cases with absent posterior horn had evidence of radiological OA (p value 0.022) even at mid-term follow-up. Conclusion: Even though the overall OAK and subjective IKDC scores did not show significant difference between the two subsets, the poorer outcomes in terms of objective stability and radiological OA noted in the absence of PHMM, indicates the importance of preserving this important part of the meniscus.

Keywords: ACL, functional outcome, knee, posterior of medial meniscus

Procedia PDF Downloads 337
24251 Bridge Members Segmentation Algorithm of Terrestrial Laser Scanner Point Clouds Using Fuzzy Clustering Method

Authors: Donghwan Lee, Gichun Cha, Jooyoung Park, Junkyeong Kim, Seunghee Park

Abstract:

3D shape models of the existing structure are required for many purposes such as safety and operation management. The traditional 3D modeling methods are based on manual or semi-automatic reconstruction from close-range images. It occasions great expense and time consuming. The Terrestrial Laser Scanner (TLS) is a common survey technique to measure quickly and accurately a 3D shape model. This TLS is used to a construction site and cultural heritage management. However there are many limits to process a TLS point cloud, because the raw point cloud is massive volume data. So the capability of carrying out useful analyses is also limited with unstructured 3-D point. Thus, segmentation becomes an essential step whenever grouping of points with common attributes is required. In this paper, members segmentation algorithm was presented to separate a raw point cloud which includes only 3D coordinates. This paper presents a clustering approach based on a fuzzy method for this objective. The Fuzzy C-Means (FCM) is reviewed and used in combination with a similarity-driven cluster merging method. It is applied to the point cloud acquired with Lecia Scan Station C10/C5 at the test bed. The test-bed was a bridge which connects between 1st and 2nd engineering building in Sungkyunkwan University in Korea. It is about 32m long and 2m wide. This bridge was used as pedestrian between two buildings. The 3D point cloud of the test-bed was constructed by a measurement of the TLS. This data was divided by segmentation algorithm for each member. Experimental analyses of the results from the proposed unsupervised segmentation process are shown to be promising. It can be processed to manage configuration each member, because of the segmentation process of point cloud.

Keywords: fuzzy c-means (FCM), point cloud, segmentation, terrestrial laser scanner (TLS)

Procedia PDF Downloads 206
24250 Control the Flow of Big Data

Authors: Shizra Waris, Saleem Akhtar

Abstract:

Big data is a research area receiving attention from academia and IT communities. In the digital world, the amounts of data produced and stored have within a short period of time. Consequently this fast increasing rate of data has created many challenges. In this paper, we use functionalism and structuralism paradigms to analyze the genesis of big data applications and its current trends. This paper presents a complete discussion on state-of-the-art big data technologies based on group and stream data processing. Moreover, strengths and weaknesses of these technologies are analyzed. This study also covers big data analytics techniques, processing methods, some reported case studies from different vendor, several open research challenges and the chances brought about by big data. The similarities and differences of these techniques and technologies based on important limitations are also investigated. Emerging technologies are suggested as a solution for big data problems.

Keywords: computer, it community, industry, big data

Procedia PDF Downloads 158
24249 Digital Material Characterization Using the Quantum Fourier Transform

Authors: Felix Givois, Nicolas R. Gauger, Matthias Kabel

Abstract:

The efficient digital material characterization is of great interest to many fields of application. It consists of the following three steps. First, a 3D reconstruction of 2D scans must be performed. Then, the resulting gray-value image of the material sample is enhanced by image processing methods. Finally, partial differential equations (PDE) are solved on the segmented image, and by averaging the resulting solutions fields, effective properties like stiffness or conductivity can be computed. Due to the high resolution of current CT images, the latter is typically performed with matrix-free solvers. Among them, a solver that uses the explicit formula of the Green-Eshelby operator in Fourier space has been proposed by Moulinec and Suquet. Its algorithmic, most complex part is the Fast Fourier Transformation (FFT). In our talk, we will discuss the potential quantum advantage that can be obtained by replacing the FFT with the Quantum Fourier Transformation (QFT). We will especially show that the data transfer for noisy intermediate-scale quantum (NISQ) devices can be improved by using appropriate boundary conditions for the PDE, which also allows using semi-classical versions of the QFT. In the end, we will compare the results of the QFT-based algorithm for simple geometries with the results of the FFT-based homogenization method.

Keywords: most likelihood amplitude estimation (MLQAE), numerical homogenization, quantum Fourier transformation (QFT), NISQ devises

Procedia PDF Downloads 45
24248 High Performance Computing and Big Data Analytics

Authors: Branci Sarra, Branci Saadia

Abstract:

Because of the multiplied data growth, many computer science tools have been developed to process and analyze these Big Data. High-performance computing architectures have been designed to meet the treatment needs of Big Data (view transaction processing standpoint, strategic, and tactical analytics). The purpose of this article is to provide a historical and global perspective on the recent trend of high-performance computing architectures especially what has a relation with Analytics and Data Mining.

Keywords: high performance computing, HPC, big data, data analysis

Procedia PDF Downloads 484
24247 A Landscape of Research Data Repositories in Re3data.org Registry: A Case Study of Indian Repositories

Authors: Prashant Shrivastava

Abstract:

The purpose of this study is to explore re3dat.org registry to identify research data repositories registration workflow process. Further objective is to depict a graph for present development of research data repositories in India. Preliminarily with an approach to understand re3data.org registry framework and schema design then further proceed to explore the status of research data repositories of India in re3data.org registry. Research data repositories are getting wider relevance due to e-research concepts. Now available registry re3data.org is a good tool for users and researchers to identify appropriate research data repositories as per their research requirements. In Indian environment, a compatible National Research Data Policy is the need of the time to boost the management of research data. Registry for Research Data Repositories is a crucial tool to discover specific information in specific domain. Also, Research Data Repositories in India have not been studied. Re3data.org registry and status of Indian research data repositories both discussed in this study.

Keywords: research data, research data repositories, research data registry, re3data.org

Procedia PDF Downloads 295
24246 A Study of Cloud Computing Solution for Transportation Big Data Processing

Authors: Ilgin Gökaşar, Saman Ghaffarian

Abstract:

The need for fast processed big data of transportation ridership (eg., smartcard data) and traffic operation (e.g., traffic detectors data) which requires a lot of computational power is incontrovertible in Intelligent Transportation Systems. Nowadays cloud computing is one of the important subjects and popular information technology solution for data processing. It enables users to process enormous measure of data without having their own particular computing power. Thus, it can also be a good selection for transportation big data processing as well. This paper intends to examine how the cloud computing can enhance transportation big data process with contrasting its advantages and disadvantages, and discussing cloud computing features.

Keywords: big data, cloud computing, Intelligent Transportation Systems, ITS, traffic data processing

Procedia PDF Downloads 422
24245 Enhanced Tensor Tomographic Reconstruction: Integrating Absorption, Refraction and Temporal Effects

Authors: Lukas Vierus, Thomas Schuster

Abstract:

A general framework is examined for dynamic tensor field tomography within an inhomogeneous medium characterized by refraction and absorption, treated as an inverse source problem concerning the associated transport equation. Guided by Fermat’s principle, the Riemannian metric within the specified domain is determined by the medium's refractive index. While considerable literature exists on the inverse problem of reconstructing a tensor field from its longitudinal ray transform within a static Euclidean environment, limited inversion formulas and algorithms are available for general Riemannian metrics and time-varying tensor fields. It is established that tensor field tomography, akin to an inverse source problem for a transport equation, persists in dynamic scenarios. Framing dynamic tensor tomography as an inverse source problem embodies a comprehensive perspective within this domain. Ensuring well-defined forward mappings necessitates establishing existence and uniqueness for the underlying transport equations. However, the bilinear forms of the associated weak formulations fail to meet the coercivity condition. Consequently, recourse to viscosity solutions is taken, demonstrating their unique existence within suitable Sobolev spaces (in the static case) and Sobolev-Bochner spaces (in the dynamic case), under a specific assumption restricting variations in the refractive index. Notably, the adjoint problem can also be reformulated as a transport equation, with analogous results regarding uniqueness. Analytical solutions are expressed as integrals over geodesics, facilitating more efficient evaluation of forward and adjoint operators compared to solving partial differential equations. Certainly, here's the revised sentence in English: Numerical experiments are conducted using a Nesterov-accelerated Landweber method, encompassing various fields, absorption coefficients, and refractive indices, thereby illustrating the enhanced reconstruction achieved through this holistic modeling approach.

Keywords: attenuated refractive dynamic ray transform of tensor fields, geodesics, transport equation, viscosity solutions

Procedia PDF Downloads 13
24244 Ordinary Differentiation Equations (ODE) Reconstruction of High-Dimensional Genetic Networks through Game Theory with Application to Dissecting Tree Salt Tolerance

Authors: Libo Jiang, Huan Li, Rongling Wu

Abstract:

Ordinary differentiation equations (ODE) have proven to be powerful for reconstructing precise and informative gene regulatory networks (GRNs) from dynamic gene expression data. However, joint modeling and analysis of all genes, essential for the systematical characterization of genetic interactions, are challenging due to high dimensionality and a complex pattern of genetic regulation including activation, repression, and antitermination. Here, we address these challenges by unifying variable selection and game theory through ODE. Each gene within a GRN is co-expressed with its partner genes in a way like a game of multiple players, each of which tends to choose an optimal strategy to maximize its “fitness” across the whole network. Based on this unifying theory, we designed and conducted a real experiment to infer salt tolerance-related GRNs for Euphrates poplar, a hero tree that can grow in the saline desert. The pattern and magnitude of interactions between several hub genes within these GRNs were found to determine the capacity of Euphrates poplar to resist to saline stress.

Keywords: gene regulatory network, ordinary differential equation, game theory, LASSO, saline resistance

Procedia PDF Downloads 612
24243 Harmonic Data Preparation for Clustering and Classification

Authors: Ali Asheibi

Abstract:

The rapid increase in the size of databases required to store power quality monitoring data has demanded new techniques for analysing and understanding the data. One suggested technique to assist in analysis is data mining. Preparing raw data to be ready for data mining exploration take up most of the effort and time spent in the whole data mining process. Clustering is an important technique in data mining and machine learning in which underlying and meaningful groups of data are discovered. Large amounts of harmonic data have been collected from an actual harmonic monitoring system in a distribution system in Australia for three years. This amount of acquired data makes it difficult to identify operational events that significantly impact the harmonics generated on the system. In this paper, harmonic data preparation processes to better understanding of the data have been presented. Underlying classes in this data has then been identified using clustering technique based on the Minimum Message Length (MML) method. The underlying operational information contained within the clusters can be rapidly visualised by the engineers. The C5.0 algorithm was used for classification and interpretation of the generated clusters.

Keywords: data mining, harmonic data, clustering, classification

Procedia PDF Downloads 218
24242 Linguistic Summarization of Structured Patent Data

Authors: E. Y. Igde, S. Aydogan, F. E. Boran, D. Akay

Abstract:

Patent data have an increasingly important role in economic growth, innovation, technical advantages and business strategies and even in countries competitions. Analyzing of patent data is crucial since patents cover large part of all technological information of the world. In this paper, we have used the linguistic summarization technique to prove the validity of the hypotheses related to patent data stated in the literature.

Keywords: data mining, fuzzy sets, linguistic summarization, patent data

Procedia PDF Downloads 245
24241 Proposal of Data Collection from Probes

Authors: M. Kebisek, L. Spendla, M. Kopcek, T. Skulavik

Abstract:

In our paper we describe the security capabilities of data collection. Data are collected with probes located in the near and distant surroundings of the company. Considering the numerous obstacles e.g. forests, hills, urban areas, the data collection is realized in several ways. The collection of data uses connection via wireless communication, LAN network, GSM network and in certain areas data are collected by using vehicles. In order to ensure the connection to the server most of the probes have ability to communicate in several ways. Collected data are archived and subsequently used in supervisory applications. To ensure the collection of the required data, it is necessary to propose algorithms that will allow the probes to select suitable communication channel.

Keywords: communication, computer network, data collection, probe

Procedia PDF Downloads 331