scholarly journals Metadata Management on Data Processing in Data Lakes

Author(s):  
Imen Megdiche ◽  
Franck Ravat ◽  
Yan Zhao
2013 ◽  
Vol 4 (1) ◽  
pp. 146-150
Author(s):  
Lax Maiah ◽  
DR.A.GOVARDHAN DR.A.GOVARDHAN ◽  
DR. C.SUNIL KUMAR

Data Warehouse (DW) is topic-oriented, integrated, static datasets which are used to support decision-making. Driven by the constraint of mass spatio-temporal data management and application, Spatio-Temporal Data Warehouse (STDW) was put forward, and many researchers scattered all over the world focused their energy on it.Although the research on STDW is going in depth , there are still many key difficulties to be solved, such as the design principle, system framework, spatio-temporal data model (STDM), spatio-temporal data process (STDP), spatial data mining (SDM) and etc. In this paper, the concept of STDW is discussed, and analyzes the organization model of spatio-temporal data. Based on the above, a framework of STDW is composed of data layer, management layer and application layer. The functions of STDW should include data analysis besides data process and data storage. When users apply certain kind of data services, STDW identifies the right data by metadata management system, then start data processing tool to form a data product which serves the data mining and OLAP. All varieties of distributed databases (DDBs) make up data sources of STDW, including Digital Elevation Model (DEM), Diagnosis-Related Group (DRG), Data Locator Group (DLG), Data Objects Management (DOM), Place Name and other databases in existence. The management layer implements heterogeneous data processing, metadata management and spatio-temporal data storage. The application layer provides data products service, multidimensional data cube, data mining tools and on-line analytical process.


1974 ◽  
Vol 13 (03) ◽  
pp. 125-140 ◽  
Author(s):  
Ch. Mellner ◽  
H. Selajstder ◽  
J. Wolodakski

The paper gives a report on the Karolinska Hospital Information System in three parts.In part I, the information problems in health care delivery are discussed and the approach to systems design at the Karolinska Hospital is reported, contrasted, with the traditional approach.In part II, the data base and the data processing system, named T1—J 5, are described.In part III, the applications of the data base and the data processing system are illustrated by a broad description of the contents and rise of the patient data base at the Karolinska Hospital.


1978 ◽  
Vol 17 (01) ◽  
pp. 36-40 ◽  
Author(s):  
J.-P. Durbec ◽  
Jaqueline Cornée ◽  
P. Berthezene

The practice of systematic examinations in hospitals and the increasing development of automatic data processing permits the storing of a great deal of information about a large number of patients belonging to different diagnosis groups.To predict or to characterize these diagnosis groups some descriptors are particularly useful, others carry no information. Data screening based on the properties of mutual information and on the log cross products ratios in contingency tables is developed. The most useful descriptors are selected. For each one the characterized groups are specified.This approach has been performed on a set of binary (presence—absence) radiological variables. Four diagnoses groups are concerned: cancer of pancreas, chronic calcifying pancreatitis, non-calcifying pancreatitis and probable pancreatitis. Only twenty of the three hundred and forty initial radiological variables are selected. The presence of each corresponding sign is associated with one or more diagnosis groups.


Metrologiya ◽  
2020 ◽  
pp. 16-24
Author(s):  
Alexandr D. Chikmarev

A single program has been developed to ensure that the final result of the data processing of the measurement calibration protocol is obtained under normal conditions. The calibration result contains a calibration function or a correction function in the form of a continuous sedate series and a calibration chart based on typical additive error probabilities. Solved the problem of the statistical treatment of the calibration protocol measuring in normal conditions within a single program “MMI–calibration 3.0” that includes identification of the calibration function in a continuous power series of indications of a measuring instrument and chart calibration. An example of solving the problem of calibration of the thermometer by the working standard of the 3rd grade with the help of the “MMI-calibration 3.0” program.


Author(s):  
D. M. Nazarov

The article describes the training methods in the course “Information Technologies” for the future bachelors of the directions “Economics”, “Management”, “Finance”, “Business Informatics”, the development of metasubject competencies of the student while his use of tools for data processing by means of the language R. The metasubject essence of the work is to update traditional economic knowledge and skills through various presentation forms of the same data sets. As part of the laboratory work described in the article, future bachelors learn to use the basic tools of the R language and acquire specific skills and abilities in R-Studio using the example of processing currency exchange data. The description of the methods is presented in the form of the traditional Key-by-Key technology, which is widely used in teaching information technologies.


2020 ◽  
Vol 6 (2) ◽  
pp. 187-197
Author(s):  
Nurlaila Suci Rahayu Rais ◽  
Dedeh Apriyani ◽  
Gito Gardjito

Monitoring of warehouse inventory data processing is an important thing for companies. PT Talaga mulya indah is still manual using paper media, causing problems that have an effect on existing information, namely: problems with data processing of incoming and outgoing goods. And the difference between data on the amount of stock of goods available with physical data, often occurs inputting data more than once for the same item, searching for available data, and making reports so that it impedes companies in monitoring inventory of existing stock of goods. Which aims to create a system that can provide updated information to facilitate the warehouse admin in making inventory reports, and reduce errors in input by means of integrated control. In this study, the authors used the data collection method used in this analysis using the method of observation, interviews, and literature review (literature study). For analysis using the PIECES analysis method. Furthermore, the system design used is UML (Unified Modeling Language). The results of this study are expected to produce the right data in the process of monitoring inventory data processing, also can provide the right information and make it easier to control the overall availability of goods.


Sign in / Sign up

Export Citation Format

Share Document