Full-text resources of CEJSH and other databases are now available in the new Library of Science.
Visit https://bibliotekanauki.pl

Results found: 11

first rewind previous Page / 1 next fast forward last

Search results

Search:
in the keywords:  data quality
help Sort By:

help Limit search:
first rewind previous Page / 1 next fast forward last
EN
This paper deals with the problem of unequal representation of countries in international surveys, and the differences in data quality between survey projects, both obstacles to cross-national comparative research. The first part of the paper investigates international surveys on general population samples conducted in South-East Europe in the period between 1990 and 2010. Documentation of country participation in both general and region- or theme-specific survey projects shows that some countries are systematically excluded from surveys. Consequently, from comparative perspective, the generalizability of research results is not only limited but also potentially biased, omitting atypical cases. The second part of the paper focuses on the quality of surveys. It finds that the most problematic element of surveys is survey documentation, an essential component of the data. Without documentation the value of datasets, analyses using them and conclusions drawn on their basis are questionable. The proposed synthetic measure of data quality, the Survey Quality Index, could lead to setting standards for the documentation of the survey process, and thus facilitate cross-national research and allow for meaningful integration of existing survey data.
EN
Data quality is crucial in today’s business processes, as it is generally associated with the set of data fit for use by data consumers - the persons that access, interpret and use data during their work activity. On the other hand, data quality is very important for the Accounting Information Systems’ (AIS) success, where AIS is a computer-based system that processes financial data and supports the decision making processes inside the organization. There are empirical evidences showing that data quality level in AIS has been and will always be problematic. Their interrelationship is dependant of several factors, including technical capacities or even the level of teamwork in an organization. This paper tries to analyze the actual performance of the factors influencing in the process of data quality in AIS used from organizations in Albania. The results will be compared with state-of-art literature review regarding the factors perceived as critical factors in ensuring data quality in AIS, giving way to some important concluding remarks.
EN
The paper is focused on the processing of 3D weather radar data to minimize the impact of a number of errors from different sources, both meteorological and non-meteorological. The data is also quantitatively characterized in terms of its quality. A set of dedicated algorithms based on analysis of the reflectivity field pattern is described. All the developed algorithms were tested on data from the Polish radar network POLRAD. Quality control plays a key role in avoiding the introduction of incorrect information into applications using radar data. One of the quality control methods is radar data assimilation in numerical weather prediction models to estimate initial conditions of the atmosphere. The study shows an experiment with quality controlled radar data assimilation in the COAMPS model using the ensemble Kalman filter technique. The analysis proved the potential of radar data for such applications; however, further investigations will be indispensable.
EN
Neonatal mortality rates by gestational age and birth weight category are important indicators of maternal and child health and care quality. However, due to recent laws on administrative simplification and privacy, these specific rates have not been calculated in Italy since 1999. The main aim of this work is to assess the possibility of retrieving information on neonatal mortality by the linkage between records related to live births and records related to infant deaths within the first month of life, with reference to 2003 and 2004 birth cohorts. From a strict methodological point of view, some critical aspects of the most used record linkage approach are highlighted: specific problems may arise from the choice of records to be linked if there are consistency constraints between pairs (in this context, one death record can be linked to at most one birth record). In the light of considerations on the quality of the starting data, the retrieval of information on neonatal mortality by gestational age and birth weight is restricted to Northern Italy. Specific neonatal mortality rates are provided with reference to 2003 and discussed with particular emphasis on quality issues in the data collection processes.
EN
The author reviews theory and application of rotation methods in sample surveys in Poland. He begins with reviewing designs of the surveys across time, depending on different objectives, focusing on partial rotation of sub-samples, and considers estimation problems and data quality issues generally. Next, he refers to some articles and books about surveys published over time, starting with Wilks (1940), Patterson (1950), Eckler (1955), Woodruff (1963), Rao and Graham (1964), Bailar (1975), Duncan and Kalton (1987) and Kalton and Citro (1993). He mentions also early Polish papers on rotation methods (Kordos (1966, 1967, 1971, 1982); Lednicki, 1982; Szarkowski and Witkowski, 1994), and concentrates on Polish household surveys, mainly Household Budget Survey (HBS), Labour Force Survey (LFS) and EU Statistics on Living Conditions and Income (EU-SILC). Special attention is devoted to last research on rotation sampling done by Polish sampling statisticians: Ciepiela et al. (2012), Kordos (2002), Kowalczyk (2002, 2003, 2004), Kowalski (2006, 2009), Kowalski and Wesołowski (2012) and Wesołowski (2010). Concluding remarks are given at the end.
EN
The array of archival maps from the 19th and 20th centuries is very large and, in order to assess their reliability for a particular analysis, some kind of evaluation form must be used. The proposed evaluation form comprises both formal and quantitative criteria, enriched with the maps’ elaboration circumstances, which may influence their reliability. These factors are also applied at both the spatial and attribute levels of information. Guidelines include: the scope of content, the map’s mathematical precision, the descriptive information correctness, the time reference of the content, and the information transfer efficiency.
EN
The article presents a conceptual model for data quality management treated as the usability or the compliance of the data product with its specification. The proposed model refers to the well-known TDQM model of Wang based on the Deming's quality improvement cycle. However, the TDQM model does not take into account the impact of the Internet environment on the quality of the data provided by the systems on the Web. The author's model presented in this article takes into account the impact of the Internet on all aspects resulting from data functions in society and organizations. Therefore, it takes into consideration the aspect of promoting data quality management processes, the communication aspect and the aspect of enrichment of individual and collective knowledge. The model also takes into account the fact that the impact of the known properties of the Internet (defined with the acronym MEDIA for example) refers primarily to the contextual quality characteristics of the data on the Web and, only to a small degree, it concerns the internal quality of information pieces described by such features as accuracy, consistency, complexity and precision.
PL
W artykule przedstawiono konceptualny model zarządzania jakością informacji traktowanej jako jej użyteczność lub zgodność produktu informacyjnego z jego specyfikacją. Proponowany model nawiązuje do znanego modelu TDQM R.Y. Wanga opartego na cyklu Deminga doskonalenia jakości. Jednakże model TDQM nie uwzględnia wpływu środowiska Internetu na jakość informacji udostępnianej przez systemy informacyjne w sieci WWW. Zaprezentowany w artykule autorski model bierze pod uwagę wpływ właściwości Internetu na wszystkie aspekty wynikające z funkcji informacji w społeczeństwie i w organizacji. Uwzględnia zatem aspekt wspierania procesów zarządzania jakością informacji, aspekt komunikacyjny oraz aspekt wzbogacania wiedzy indywidualnej i zbiorowej. W modelu uwzględniono także fakt, że wpływ znanych właściwości Internetu (określonych np. akronimem MEDIUM) odnosi się przede wszystkim do kontekstowych cech jakości informacji w sieci WWW, a w małym stopniu dotyczy wewnętrznej jakości jednostek informacji opisanych takimi cechami jak np. dokładność, spójność, złożoność czy precyzja.
EN
The goal of this article is to inform social scientists, especially those of a quantitative orientation, about the basic characteristics of Big Data and to present the opportunities and limitations of using such data in social research. The paper informs about three basic types of Big Data as they are distinguished in contemporary methodological literature, namely administrative data, transaction data and social network data, and exemplifies how they can be utilized by quantitative social research. According to many, questionnaire-based sample survey as the dominant method of quantitative social research has found itself in a crisis, especially as response rates have decreased in most developed countries and public confidence in opinion polling has declined. The author presents the characteristics and specifics of Big Data compared to survey research - a method whose primary distinguishing characteristic is the capacity to quantify individual behaviour, social action and attitudes at the level of populations. In this context, the article draws attention to the differences between Big Data and survey data typically presented in scholarly literature, namely that datasets are not representative of known populations, the values of observed variables are systematically biased, there is a limited number of variables in Big Data sets, there is uncertainty about the meaning of observed values, and social environment has direct influence on the behaviours captured by Big Data. Attention is also paid to such characteristics of Big Data that pose an obstacle to smooth integration of this type of data in the social scientific mainstream. First, the collection, processing and analysis of Big Data is extremely demanding in terms of programming skills, something social scientists typically do not have. Second, the availability of Big Data is limited as they are normally possessed by private corporations, some of which (Facebook, Google) have undoubtedly come to form data oligopolies - and their management is mostly unwilling to share their data with traditional academics. Based on the above-mentioned specifics, differences and limitations, it is argued that Big Data currently do not have the potential of becoming a full-fledged source of social science data and replacing sample surveys as the dominant research method. Finally, the article draws attention to the specifics of different types of Big Data as they are primarily generated for purposes other than social research and result from specific situations framed by existing social relations - and it is from this perspective that Big Data should be viewed by social researchers.
EN
The author begins with a general assessment of the mission of the National Statistics Institutes (NSIs), main producers of official statistics, which are obliged to deliver high quality statistical information on the state and evolution of the population, the economy, the society and the environment. These statistical results must be based on scientific principles and methods. They must be made available to the public, politics, economy and research for decision-making and information purposes. Next, before discussing general issues of small area estimation (SAE) in official statistics, the author reminds: the methods of sampling surveys, data collection, estimation procedures, and data quality assessment used for official statistics. Statistical information is published in different breakdowns with stable or even decreasing budget while being legally bound to control the response burden. Special attention is paid, from a practitioner point of view, to synthetic development of small area estimation in official statistics, beginning with international seminars and conferences devoted to SAE procedures and methods (starting with the Canadian symposium, 1985, and the Warsaw conference, 1992, to the Poznan conference, Poland, 2014), and some international projects (EURAREA, SAMPLE, BIAS, AMELI, ESSnet). Next, some aspects of development of SAE in official statistics are discussed. At the end some conclusions regarding quality of SAE procedures are considered.
EN
In the research carried out to date by the authors of the article, the assessment of the quality of mirror data in the exchange of goods between European Union (EU) countries was based on the value of goods. A similar approach is applied by many researchers. The aim of the research discussed in the article is to assess the quality of data relating to intra-EU trade based on not only the value, but also on the quantity of goods. The analysis of discrepancies in data relating to trade between EU countries, with a particular emphasis on Poland, was based on selected research methods known from literature. Both the value-based and the quantitative approach constitute the authors' contribution to the development of research methodology. Data quality indicators were proposed and data pertaining to the weight of goods were used. Information on trade in goods between EU countries in 2017 obtained from Eurostat's Comext database was analysed. The research relating to the dynamics also covered the years 2005, 2008, 2011, and 2014. The results of the analysis demonstrated that the total share of export of goods from Poland to a given country within the EU is different for data expressed in value (value of goods) and in quantity (weight of goods). Therefore, both approaches should be applied in the study of the quality of mirror data.
PL
W badaniach prowadzonych dotychczas przez autorów artykułu ocena jakości danych lustrzanych w wymianie towarowej między krajami Unii Europejskiej (UE) opierała się na wartości towarów. Analogiczne podejście stosuje wielu badaczy. Celem badania omawianego w artykule jest ocena jakości danych dotyczących obrotu wewnątrzunijnego na podstawie nie tylko wartości, lecz także ilości towarów. W analizie rozbieżności danych w handlu między krajami UE, ze szczególnym uwzględnieniem Polski, wzorowano się na wybranych metodach badania znanych z literatury przedmiotu. Podejście zarówno wartościowe, jak i ilościowe stanowi wkład własny autorów w metodykę badawczą. Zaproponowano wskaźniki jakości danych oraz wykorzystano dane dotyczące masy towarów. Analizie poddano informacje na temat obrotu towarowego między krajami unijnymi w 2017 r. pochodzące z bazy Eurostatu Comext. Badanie dynamiki obejmowało również lata: 2005, 2008, 2011 i 2014. Wyniki analizy pokazały, że ogółem udział wywozu z Polski towarów do danego kraju na obszarze UE jest różny dla danych wyrażonych wartościowo (wartość towarów) i ilościowo (masa towarów). W badaniu jakości danych lustrzanych należy zatem stosować oba podejścia.
PL
Autor bada współzależności między rozwojem teorii i praktyki badań reprezentacyjnych w Polsce w ciągu ponad 60 lat. Rozpoczyna od klasycznej pracy Neymana (Neyman, 1934), która dała teoretyczne podstawy probabilistycznego podejścia wyboru próbki umożliwiające wnioskowanie z badań reprezentacyjnych. Główne idee tej pracy były najpierw opublikowane po polsku w 1933 r. (Neyman, 1933) i miały istotny wpływ na a praktykę badań reprezentacyjnych Polsce przed i po II Wojnie Światowej. Badania reprezentacyjne prowadzone w latach 1950-tych i 1960-tych były konsultowane z J. Neymanem w czasie jego wiz w Polsce w latach 1950 i 1958 (Fisz, 1950; Zasępa, 1958). Praktyczne problemy występujące przy planowaniu i analizie badań reprezentacyjnych Polsce były częściowo rozwiązywane przez Komisję Matematyczną GUS powołaną w końcu 1949 r. jako organ doradczy i opiniodawczy Prezesa GUS. Komisja ta składała się ze specjalistów zarówno z GUS jak i ośrodków naukowo-badawczych w kraju (Kordos, 2012a). Komisja działała do 1993 r. i wpływała w istotny sposób na praktykę badań reprezentacyjnych Polsce. Specjalną uwagę poświęcono wpływowi teorii badań próbkowych na świecie i w Polsce na praktykę badań reprezentacyjnych w Polsce, a w szczególności na plany i schematy losowania i metody estymacji w badaniach prowadzonych w czasie, a głównie metodzie rotacyjnej (Greń, 1969; Kordos, 1967, 2012b; Kowalczyk, 2004; Kowalski, 2006; Lednicki, 1982; Popiński, 2006; Wesołowski, 2010); lokalizacji próby (Bracha et al.,2004b, Greń, 1964, 1966; Lednicki, 1979, 1989; Lednicki i Wesołowski, 1994;), metodom estymacji (Bracha, 1996, 1998; Greń, 1970; Kordos, 1982; Lednicki, 1979, 1987, Wesołowski, 2004, Zasępa, 1962, 1972), jakości danych (Kordos, 1973, 1988; Zasępa, 1993) oraz metodom estymacji dla małych obszarów (Bracha, 1994, 2003; Bracha et al., 2004b; Dehnel, 2010; Domański, Pruska, 2001; Golata, 2004ab, 2012; Kalton et al., 1993; Kordos, 1991, 2000b; Kordos, Paradysz, 2000; Niemiro, Wesołowski, 2012; Paradysz, 1998; Wesołowski, 2004). W zakończeniu prowadzone są rozważania na temat przyszłych zapotrzebowań na informacje z badań reprezentacyjnych.
EN
The author examines interplay between sample survey theory and practice in Poland over the past 60 years or so. He begins with the Neymans’s (1934) classic landmark paper which laid theoretical foundations to the probability sampling (or design—based) approach to inference from survey samples. Main ideas of that paper were first published in Polish in 1933 (Neyman, 1933) and had a significant impact on sampling practice in Poland before and after the World War II. Sample surveys conducted in 1950s and 1960s were consulted with J. Neyman during his visits in Poland in 1950 and 1958 (Fisz, 1950a; Zasepa, 1958). Some practical problems encountered in the design and analysis of sample surveys were partly solved by the Mathematical Commission of the CSO which was established in 1949, as an advisory and opinion—making body to the CSO President in the field of sample surveys. The Commission concentrated specialists in the sampling methods both from the CSO and research centres in the country (Kordos, 2012a). The Commission had a significant impact on sampling practice in Poland and was active till 1993. Special attention is devoted to the impact of sampling theory on sampling practice in Poland, and particularly on sample designs and estimation methods in: sampling in time and rotation methods (Kordos, 1967, 2012b; Kowalczyk, 2004; Kowalski, 2006; Wesołowski, 2010); sample allocation and estimation methods (Bracha, 1994, 1996, 2003; Greń, 1964, 1966, 1969, 1970; Kordos, 1969, 1973, 1982; Wesołowski, 2004, Zasępa, 1962, 1972, 1993 ); data quality (Kordos, 1973, 1988); estimation methods for small areas (Dehnel, 2010; Domański, Pruska, 2001; Golata, 2004ab, 2012; Kalton et al., 1993; Kordos, 1991, 2000b, 2004; Kordos, Paradysz, 2000; Niemiro, Wesołowski, 2012; Paradysz, 1998; Wesołowski, 2004). Concluding remarks are given at the end.
first rewind previous Page / 1 next fast forward last
JavaScript is turned off in your web browser. Turn it on to take full advantage of this site, then refresh the page.