Abstract
PurposeThe purpose of this paper is to compare two ways of collecting job quality data in Britain using a common set of questions. One way is through a short quiz taken by a self-selected sample and completed by clicking on a web link www.howgoodismyjob.com. The other way is via an invitation to take part in a long-running survey of working life – the Skills and Employment Survey. The survey takes much longer to complete, is carried out face-to-face and is based on random probability principles.Design/methodology/approachTo be content-comparable, the quiz uses tried and tested questions contained in recent waves of the Skills and Employment Survey. Each survey comprises a nationally representative sample of workers in Britain aged 20–65 years. However, the quiz is based on uncontrolled convenience sampling prompted, in large part, by a Facebook advertising campaign, whereas survey participants are randomly selected. In this paper, the authors compare the profile of respondents and their responses to these two different modes of data collection and therefore shine a light on any biases in the samples and differences in the results respondents report.FindingsThe paper shows that while the number taking in the quiz is impressive, participation in the quiz – unlike the survey – is heavily skewed. Weighting can be used to correct some of these sample selection biases. But, even then, the picture painted by the quiz and survey data varies with the quiz under-reporting the intrinsic quality of jobs, while over-reporting on the extrinsic rewards. This suggests that how job quality data are collected can have a strong influence on the results produced.Research limitations/implicationsThe findings suggest that a number of biases are in operation, both in terms of those who take part and the answers they give. This makes comparison between data collected using radically different methods, at best, inadvisable and, at worst, misleading. Nevertheless, quizzes are a good way of engaging large numbers of people in public debates, gathering additional data, extending the reach of academic work and prompting action to improve working life. However, the limitation of this study is that it does not offer a true experiment of different ways of collecting the same data. The quiz and survey were, for example, not carried out at the same time, but were some 14 months apart.Practical implicationsOver 50,000 people took part in the quizzes reported in the paper and almost 1,300 investigated joining a trade union as a result. The reach of the quiz far exceeds the 3,306 people who took part in the Skills and Employment Survey 2017.Originality/valueThis paper focusses on how job quality data are collected and the consequences this has for the validity of the data gathered. This is a unique contribution to international debates about the measurement and monitoring of trends in job quality.
Subject
Organizational Behavior and Human Resource Management,Industrial relations
Reference30 articles.
1. Job demands-resources model: state of the art;Journal of Managerial Psychology,2007
2. Mode of questionnaire administration can have serious effects on data quality;Journal of Public Health,2005
3. DeMaio, T.J. (1984), “Social desirability and survey measurement: a review”, in Turner, C.F. and Martin, E. (Eds), Surveying Subjective Phenomena, Russell Sage, New York.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献