The Role of Data Analysis in Academic Research: Best Practices

Data analysis is a critical component of academic research that allows researchers to transform raw data into meaningful insights. By applying appropriate analysis techniques, researchers can uncover patterns, relationships, and trends that help support their research questions and hypotheses. In academic research, data analysis not only helps in drawing conclusions but also ensures that the research findings are credible, reliable, and valid. In this article, we’ll explore the importance of data analysis in academic research and the best practices that researchers should follow to ensure accurate and effective data interpretation.

1. Understanding the Importance of Data Analysis

Data analysis serves as the backbone of most academic research. It helps researchers make sense of large volumes of raw data, enabling them to identify significant patterns and relationships. Effective data analysis leads to the development of strong, evidence-based conclusions that support or refute research hypotheses. Whether you are conducting qualitative or quantitative research, data analysis is essential for transforming data into actionable knowledge.

Why it’s important: Data analysis ensures that the research findings are based on solid, objective evidence, contributing to the credibility of the study.

Tip: Approach data analysis with an open mind, ready to discover unexpected trends and patterns that may alter your conclusions.

2. Choosing the Right Data Analysis Techniques

Different types of data require different analysis methods. Researchers must choose the right data analysis technique based on their research questions and the nature of the data they’ve collected. For quantitative research, techniques such as regression analysis, correlation analysis, and hypothesis testing are commonly used. For qualitative research, methods like thematic analysis, content analysis, and coding are often employed.

Why it’s important: Selecting the appropriate data analysis technique ensures that the results of your analysis are meaningful and accurate.

Tip: Familiarize yourself with various data analysis methods and consult with statistical or data analysis experts if needed to choose the best approach for your study.

3. Ensuring Data Accuracy and Quality

Before analyzing data, it is crucial to ensure that it is accurate and of high quality. Data accuracy can be compromised by errors during data collection, coding, or entry. Researchers must take steps to clean and validate their data to eliminate inaccuracies. This process includes checking for missing values, outliers, and inconsistencies that could affect the results.

Why it’s important: High-quality, accurate data ensures that the conclusions drawn from your research are reliable and trustworthy.

Tip: Regularly check and clean your data throughout the research process to prevent issues from compounding later in your analysis.

4. Using Statistical Software for Efficient Data Analysis

Modern research heavily relies on statistical software tools to analyze large datasets efficiently. Tools like SPSS, R, SAS, and Python offer powerful functions for data analysis, helping researchers perform complex statistical operations, visualize data trends, and generate accurate results. These tools can streamline the data analysis process, saving time and reducing the risk of errors.

Why it’s important: Statistical software helps researchers conduct sophisticated analyses and process large volumes of data quickly and accurately.

Tip: Learn how to use the relevant statistical software for your research, and take advantage of online tutorials and resources to improve your skills.

5. Interpreting Data Correctly

Interpreting data is just as important as analyzing it. After performing the data analysis, researchers must interpret the results in the context of their research questions and objectives. This involves identifying patterns, drawing comparisons, and determining the implications of the findings. Researchers should also be cautious of biases that may influence their interpretation and ensure that conclusions are supported by the data.

Why it’s important: Correct interpretation of data is critical for making valid conclusions and contributing to the overall academic discourse.

Tip: Review your data interpretation in light of your research hypothesis and consult with peers or mentors to ensure the findings align with the data.

For more tips on academic research, book free consultation with our expert writers today!

6. Presenting Data and Findings Effectively

Presenting data in a clear and comprehensible manner is key to conveying the significance of your research findings. Use graphs, charts, tables, and visual aids to display data in a way that is easy to understand. A well-organized presentation allows readers to quickly grasp your results and makes it easier for them to follow your argument.

Why it’s important: Effective presentation of data makes your research findings more accessible to a wider audience and enhances the impact of your work.

Tip: Use visuals sparingly and ensure that each chart or graph is properly labeled with clear titles, axes, and legends to avoid confusion.

7. Avoiding Common Data Analysis Pitfalls

There are several common pitfalls in data analysis that researchers should be aware of, such as overfitting, underfitting, or misinterpreting correlation as causation. Researchers may also be tempted to manipulate data to fit a desired outcome, which can lead to biased results and ethical issues. To avoid these pitfalls, researchers should remain objective and follow established data analysis protocols.

Why it’s important: Avoiding common pitfalls ensures the integrity of your research and the validity of your conclusions.

Tip: Regularly review your analysis methodology and remain open to alternative interpretations of the data.

Data analysis is a fundamental part of academic research, as it transforms raw data into meaningful insights that support research conclusions. By understanding the importance of data analysis, choosing the right techniques, ensuring data quality, and presenting findings effectively, researchers can produce high-quality work that contributes valuable knowledge to their field. Following best practices in data analysis not only enhances the credibility of your research but also increases the likelihood of making a meaningful impact in your academic community.

If you’re ready to improve your data analysis skills, follow these best practices and apply them to your next research project. For more tips on academic research, book free consultation with our expert writers today! [email protected]

All rights reserved © 2024 • Powered by 99Scholar

WhatsApp us

null

Enter the URL below into your favorite RSS reader.

Approaches to Analysis of Qualitative Research Data: A Reflection on the Manual and Technological Approaches

  • Citation (BibTeX)

data analysis in research google scholar

Sorry, something went wrong. Please try again.

If this problem reoccurs, please contact Scholastica Support

Error message:

View more stats

This paper addresses a gap in the literature by providing reflective and critical insights into the experiences during two PhD qualitative studies, which adopted different approaches to data analysis. We first consider how the two PhD studies unfolded before discussing the motivations, challenges and benefits of choosing either a technological (NVivo) or manual approach to qualitative data analysis. The paper contributes to the limited literature which has explored the comparative experiences of those undertaking qualitative data analysis using different approaches. It provides insights into how researchers conduct qualitative data analysis, using different approaches and the lessons learnt.

1. Introduction

Qualitative data analysis has a long history in the social sciences. Reflecting this, a substantial literature has developed to guide the researcher through the process of qualitative data analysis (e.g. Bryman & Burgess , 1994; Harding , 2018; Saunders et al. , 2019; Silverman , 2017 ). While earlier literature focuses on the manual approach [1] to qualitative data analysis (Bogdan & Bilken , 1982; Lofland , 1971) , more recent literature provides support in the application of a range of technological approaches (alternatively referred to as Computer Assisted Qualitative Data Analysis Software or CAQDAS): e.g., Excel (Meyer & Avery , 2009) ; NVivo (Jackson & Bazeley , 2019) ; and ATLAS.ti (Friese , 2019) . Moreover, in an accounting context, a critical literature has emerged which attempts to elucidate the messy and problematic nature of qualitative data analysis (Ahrens & Chapman , 2006; Lee & Humphrey , 2006; Modell & Humphrey , 2008; O’Dwyer , 2004; Parker , 2003) . However, while a substantial literature exists to guide the researcher in undertaking qualitative data analysis and in providing an understanding of the problematic nature of such analyses, a dearth of research reports on the comparative experiences of those undertaking qualitative data analysis using different approaches. The paper aims to address this gap by reporting on the experiences of two recently qualified doctoral students as they reflect on how they each approached the task of analysing qualitative data, Researcher A (second author) choosing a technological approach (NVivo) while Researcher B (third author) opted for a manual approach. The paper contributes to the limited literature which explores the comparative experiences of those undertaking qualitative data analysis using different approaches. In so doing, we hope that the critical reflections and insights provided will assist qualitative researchers in making important decisions around their approach to data analysis.

The remainder of the paper is structured as follows. In section two, we provide an overview of the problematic nature of qualitative research and a review of the manual and technological approaches of data analysis available to researchers. Section three follows with a discussion of two qualitative PhD studies. Section four discusses the experiences, challenges and critical reflections of Researchers A and B as they engaged with their particular approach to qualitative data analysis. The paper concludes with a comparative analysis of the experiences of Researchers A and B and implications for further work.

2. Literature Review

2.1 a qualitative research approach: debates and challenging issues.

Qualitative researchers pursue qualia , that is phenomena as experienced (sometimes uniquely) by individuals, that enlarge our conception of the “really real” (Sherry & Kozinets , 2001 , p. 2) . Qualitative studies seek to answer ‘how’ and ‘why’ rather than ‘what’ or ‘how often’ questions. In so doing, qualitative studies involve collecting rich data that are understood within context and are associated with an interpretivist philosophy. Mason (2002) notes that qualitative research is not just about words, rather it reflects a view of practice that is socially constructed and requires researchers to embrace subjectivity in order to interpret data. Furthermore, Bédard & Gendron (2004) argue that “being tolerant of uncertainty is part of the fundamental skills of the qualitative researcher” (p. 199). That said, a qualitative approach can be extremely labour intensive, given the volume of data collected and the commitment required to generate themes.

In the accounting and management literatures, there has been considerable debate on the challenges of qualitative data analysis. In early work, Parker (2003) highlights a potential challenge in that qualitative researchers need to be reflexive in the data analysis process. To that end, researchers often construct field notes and memos (during interviews for example) to report their feelings, perceptions and impressions, which can be viewed as data, alongside all other data collected from the field. Bédard & Gendron (2004) highlight a further challenge in that analysing qualitative data is both labour intensive and requires high levels of research knowledge and ability. Furthermore, they argue that qualitative researchers need to be immersed in data collection and analysis, and should be mindful that the “specific objectives of the study are not always determined a priori, but often ‘emerge’ from fieldwork” (p. 200). Ahrens & Chapman (2006) identify the challenge of data reduction without “‘thinning’ out the data to the point where it loses its specificity and becomes bland” (p. 832). Qualitative data analysis is, they argue, not a straightforward process: “Like other practices, the doing of qualitative field studies is difficult to articulate. One can point to the golden rules but, at the heart of it lies a problem of transformation. Out of data, snippets of conversations and formal interviews, hours and days of observation, tabulations of behaviours and other occurrences, must arise the plausible field study” (Ahrens & Chapman , 2006 , p. 837) . This chimes with O’Dwyer’s (2004) description of qualitative data analysis as ‘messy’. To address this, O’Dwyer (2004) highlights the importance of imposing structure onto the analysis process and outlines an intuitive approach to analyse interview data using Miles and Huberman’s (1994) three stage process of data reduction, data display and data interpretation/conclusion drawing and verification. This process involves the categorisation of themes and individual aspects of interviews in several stages to ensure that general patterns and differences are articulated. While O’Dwyer (2004) considered using a technological approach to assist in data analysis, he discounted it as an option at an early stage of his research, largely as a result of his lack of understanding of what it could offer. Lee & Humphrey (2006) also argue that analysing interview transcripts is a key challenge facing qualitative researchers. In particular, deciding “what weight to give to meanings that are only apparent in a part of an interview, how to retain understanding of the whole interview when the focus is on individual parts and how to derive patterns both within and across interviews without losing sight of any idiosyncratic elements that may provide unique insights” (p. 188). Finally, Modell & Humphrey (2008 , p. 96) , while calling for further research in the area of qualitative data analysis, contend that problems exist where there is undue focus on the approach to data analysis to the detriment of the development of ideas. They suggest that this appears to be an increasingly common issue, particularly with increased use of technology in the data analysis process.

2.2 Approaches to Data Analysis: Manual and Technological (i.e. NVivo) Approaches

The data analysis phase of qualitative research is described as the “most intellectually challenging phase” (Marshall & Rossman , 1995 , p. 114) and the active role of the researcher in identifying and communicating themes is critical (Braun & Clarke , 2006; Edwards & Skinner , 2009; Silverman , 2017) . While early technological approaches to data analysis have been in existence since the 1960s, many qualitative researchers have continued to employ the manual approach to analysis (Séror , 2005) . In part, this may be due to the perceptions of some researchers that the technological approach may attempt to do more than assist in the management of data, potentially influencing the abstraction of themes from data in unintended ways (Crowley et al. , 2002) . However, a review of the literature suggests that the manual approach can be an unwieldy, cumbersome, “tedious and frustrating” process (Basit , 2003 , p. 152) . Furthermore, comparatively little has been published in relation to the mechanics of the manual approach (Bazeley , 2009; Bogdan & Bilken , 1982; Braun & Clarke , 2006; Edwards & Skinner , 2009; Lofland , 1971; Maher et al. , 2018; Miles & Huberman , 1994; Silverman , 2017) .

Edwards & Skinner (2009) assert that the manual analysis of hundreds of pages of raw data is a “daunting” task (p. 134). To assist in this process, some basic mechanical procedures are described in the literature, including: printing hardcopy transcripts, photocopying, marking up, line-by-line coding, coding in margins, cutting, cut-and-paste, sorting, reorganising, hanging files and arranging colour-coded sticky notes on large format display boards (Basit , 2003; Bogdan & Bilken , 1982; Lofland , 1971; Maher et al. , 2018; L. Richards & Richards , 1994) . Moreover, Braun & Clarke (2006) provide a comprehensive description of the manual data analysis process, involving “writing notes on the texts you are analysing, by using highlighters or coloured pens to indicate potential patterns, or by using ‘post-it’ notes to identify segments of data” (p. 89). As ‘codes’ are identified, data extracts are manually grouped and collated within the individual codes. The subsequent generation of sub-themes and overarching themes involves the trialling of combinations of codes until “all extracts of data have been coded in relation to them” (p. 89). The above is an iterative process and involves re-reading, coding and recoding until all data has been included in sub-themes and overarching themes. The researcher’s interaction with the data is important in this regard, and involves a series of physical activities around arranging and re-arranging data excerpts and post-it notes, followed by visual mapping on “large format display boards” (Maher et al. , 2018 , p. 11) . This process “encourages a slower and more meaningful interaction with the data [and] great freedom in terms of constant comparison, trialling arrangements, viewing perspectives, reflection and ultimately developing interpretative insights” (Maher et al. , 2018 , p. 11) .

An alternative to the manual approach is the use of CAQDAS (i.e. technological approach) to support qualitative data analysis. CAQDAS offers the ability to import, organise and explore data from various sources (text, audio, video, emails, images, spreadsheets, online surveys, social and web content). The origins of NVivo, one of the market leaders, can be traced back to the 1980s with the development of a computer programme called Non-numerical Unstructured Data Indexing Searching and Theorizing (NUD*IST). Richards, one of the co-developers of NVivo provides an “intellectual history” of NUD*IST and NVivo (R. Richards , 2002 , p. 199) , arguing that “NVivo … is being preferred by researchers wishing to do a very detailed and finely articulated study … [and that NVivo’s] tools support close and multi-faceted analysis on small and moderate amounts of data” (p. 211). Reflecting its widespread usage as a mainstream CAQDAS, a literature has now developed around the use of NVivo. For example, Bandara (2006) provides guidance to novice researchers and academics involved in NVivo research training in information systems research; García-Horta & Guerra-Ramos (2009) provide reflections on the use of NVivo in education; Leech & Onwuegbuzie (2011) present guidance for psychology researchers; and Zamawe (2015) presents experiences in the context of health professionals.

Acknowledging that little is known about how researchers use CAQDAS, Paulus et al. (2017) present the results of a discourse analysis of some 763 empirical studies which use NVivo or ATLAS.ti (a competitor of NVivo – see https://atlasti.com/ ). Drawing on peer reviewed papers, published between 1994 and 2013, Paulus et al. (2017) report that the majority of researchers (87.5% of their sample) using CAQDAS to support qualitative data analysis fail to provide details of the technological approach used beyond naming the software, or what they refer to as ‘name-dropping’. Some 10% of the sample provide moderate levels of reporting, mainly concerned with “descriptions of software capability” (Paulus et al. , 2017 , p. 37) . The remaining 2% of the sample provide more detailed descriptions of the CAQDAS used, including “detailed descriptions of how the analysis was conducted” (p. 39) or “how the researchers used the software to go beyond coding to a deeper layer of analysis” (p. 41). Based on their findings, Paulus et al. (2017) suggest that future studies should provide more detail about their experiences of using CAQDAS to support qualitative data analysis, including: what software is used; how they are used; why they are used; and how effective they have been.

A limited number of studies report on the benefits and drawbacks of using NVivo. In an early study, García-Horta & Guerra-Ramos (2009) report their experiences of using NVivo (and MAX QDA ) to analyse qualitative data collected from teachers. Their experiences suggest a number of advantages, including the ability to: organise and store large volumes of data; deal with data overload; and enable fast and efficient retrieval of relevant information. However, they also highlight a number of limitations, most notably the “real hard work” of “generating categories or taxonomies, assigning meaning, synthesizing or theorizing” (p. 163) which, they argue, remains that of the researcher and not the software. García-Horta & Guerra-Ramos (2009) also highlight the potential for “data fetishism … or the ‘let’s code everything’ strategy [which] can lead to excessive and non-reflexive coding” (p. 163). They caution against the possibility of assumptions that ‘meaning-making’ can be computerised and the possibility of what they call ‘technologism’ whereby there is an implicit assumption that the qualitative data analysis process will be enhanced by the use of software. More recently, Zamawe (2015) argues that NVivo works well with most research designs as it is not methodologically specific and “the presence of NVivo makes it more compatible with grounded theory and thematic analysis approaches” (p. 14). Furthermore, Zamawe (2015) suggests NVivo eases the burden associated with manual qualitative data analysis in terms of the ‘copy-cut-paste’ requirement. NVivo also lends itself to more effective and efficient coding, and the reshaping and reorganisation of the coding structure by “simply clicking a few buttons” (p. 14). Zamawe (2015) , however, points out some pitfalls associated with using NVivo. These include: the time consuming, and difficult, nature of the software; the potential for NVivo to “take over the analysis process from the researcher” (p. 15); the process of coding the data; and the danger of the researcher becoming distant from his/her data with the result that the ‘thickness’ of the data is diluted.

2.3 Comparison of Manual and Technological Approaches

Few studies report on comparisons of the manual and technological approaches to qualitative data analysis. In one such study, Basit (2003) compares the use of the manual and technological approach to qualitative data analysis drawing on two research projects. She argues that the approach chosen is dependent on the size of the project, the funds and time available, and the inclination and expertise of the researcher. Basit (2003) maintains that while the technological approach may not be considered feasible to code a small number of interviews, it is more worthwhile when a large number of interviews are involved. When compared to the manual approach, she highlights a number of perceived benefits of the technological approach. First, the data analysis process is relatively smooth and facilitates a more in-depth analysis. Second, the search facility is particularly useful, as is the ability to generate reports. Despite the perceived benefits, Basit (2003) acknowledges some challenges of the technological approach when compared to the manual approach. There is a considerable amount of time and formal training involved in getting acquainted with a software package to code qualitative data electronically, an investment not required for the manual approach. However, that said, Basit notes that the benefit of the software search facility and the generation of comprehensive reports compensates for the time investment required. In another study, Maher et al. (2018) argue that qualitative data analysis software packages, such as NVivo, do not fully scaffold the data analysis process. They therefore advocate for the use of manual coding (such as using coloured pens, paper, and sticky notes) to be combined with digital software to overcome this. Reflecting on their research, which combined both a manual and software analysis, they argue that NVivo provides excellent data management and retrieval facilities to generate answers to complex questions that support analysis and write-up, a facility not available with a manual approach. However, they suggest that the manual approach of physically writing on sticky notes, arranging and rearranging them and visual mapping, encourages more meaningful interaction with the data, compared to a technological approach. Furthermore, they advocate that the manual approach has a particular advantage over the technological approach as manual analysis usually results in displays of the analysis. The resulting visualisations, sticky notes, and concept maps may remain in place, allowing the researcher to engage with the research material on a variety of levels and over a period of time. In contrast to the manual approach, Maher et al. (2018) believe that NVivo operated on a computer screen does not facilitate broad overviews of the data and that data views may therefore become fragmented.

The above review indicates that limited research has reported on the comparative experiences of those undertaking qualitative data analysis. This paper addresses this gap, and in so doing, reports on the experiences of two recently qualified doctoral students, as they each reflect on how they approached the task of analysing qualitative data using different approaches. Section three presents details of the two research projects.

3. The Doctoral Research Projects

In this section, the background, motivation and research question/objectives of the research projects undertaken by Researchers A and B (both undertaking a part-time PhD) are outlined. This provides context for a comparison of the technological (NVivo) and manual approaches used for qualitative data analysis.

3.1 Researcher A: Background, Motivation, Research Question and Objectives

Researcher A (a Chartered Accountant) investigated financial management practices in agriculture by exploring the financial decision-making process of Irish farmers. When the literature in the area of farm financial management (FFM) was explored, it became apparent that there were relatively few prior studies, both internationally and in the Irish context (Argiles & Slof , 2001; Jack , 2005) . The limited literature posed particular difficulties and frustration when conducting this research, but also demonstrated that there was a gap in the literature that needed to be addressed. The review of the literature identified a number of key issues which were central to the motivation of the research project. First, the majority of farmers appear to spend very little time on financial management (Boyle , 2012; Jack , 2005) and second, farmers tend to rely on intuition to a large extent when managing their farm enterprise (Nuthall , 2012; Öhlmér & Lönnstedt , 2004) .

Researcher A’s overall research question was: How and why do farmers make financial decisions? To address this question, two research objectives were formulated following a detailed literature review and findings from preliminary research, namely a pilot survey of farmers and key informant interviews. The theoretical framework adopted (sensemaking theory) also assisted in framing the research objectives.

Research Objective 1: To explore the financial decision-making process of farmers by examining:

The factors that influence farmer decision-making;

The role of advisors in farmer decision-making;

The role of FFM in farmer decision-making;

The role of other issues in farmer decision-making (e.g. demographic factors such as farm type, age and level of education of the farmer, and the role of intuition in farmer decision-making).

Research Objective 2: To establish how farmers make sense of their business situations in order to progress with decisions of a financial nature.

The research methodology chosen by Researcher A was interpretivist in nature (Ahrens & Chapman , 2006) . This was based on the assumption that farmers’ realities (in regard to how financial decisions are made) are subjective, socially constructed and may change. As a result, it was considered necessary to explore the subjective meanings motivating the decisions of farmers in order to understand the farmers’ decision-making processes. Interviews were considered the most appropriate data collection method to operationalise the interpretivist methodology chosen. The data collected via interviews with farmers allowed Researcher A to develop thick and rich explanations of how farmers make financial decisions.

3.2 Researcher B: Background, Motivation, Research Question and Objectives

Researcher B (also a Chartered Accountant) examined accounting practitioners’ perceptions of professional competence and their engagement with Continuing Professional Development (CPD) activities, as they strive to maintain and develop competence. Educational guidance on mandatory CPD within the profession was introduced in 2004 (IES 7 , 2004) , and while CPD is viewed as a bona fide stage in the lifecycle of professional education, it is in a state of infancy and transition and has yet to grow to achieve coherence, size and stature equivalent to the pre-qualification stage (Friedman & Phillips , 2004) . While professional accountancy bodies may interpret IES 7 guidance and almost exclusively decide what counts as legitimate or valid CPD, individual practitioners are mandated to complete and self-certify relevant activities on an annual basis in order to retain professional association. It is therefore questionable whether the annual declaration encapsulates the totality of practitioners’ learning and professional development in relation to professional competence (Lindsay , 2013) .

A review uncovered an extensive literature, concentrating on professionalisation, competence and professional education and learning, with attention focusing on the accounting domain. The following emerged: literature on professionalisation pertaining to the pre-qualification period (Flood & Wilson , 2009) ; findings on competence, education and learning largely focusing on higher education (Byrne & Flood , 2004; Paisey & Paisey , 2010) ; and CPD studies predominantly reporting on engagement (Paisey et al. , 2007) . The literature review highlighted a research gap and acknowledged the need for enhanced understanding in relation to post-qualification stages, where learning and professional development could more appropriately be examined from a competence angle (Lindsay , 2013) .

The overall research objective of Researcher B’s study was to explore how individual accounting professionals perceive professional competence, and how, in light of these perceptions, they manage their CPD with the purpose of maintaining and further developing their professional competence. Given that the study set out to gain an understanding of individual perceptions and practices, this supported the use of an interpretivist approach (Silverman , 2017) . A phenomenographic approach (a distinct research perspective located within the broad interpretivist paradigm) was selected. The root of phenomenography, phenomenon , means “to make manifest” or “to bring light” (Larsson & Holmström , 2007 , p. 55) and phenomenography examines phenomena “as they appear to people” (Larsson & Holmström , 2007 , p. 62) . The phenomenographic approach is an experiential, relational and qualitative approach, enabling the researcher to describe the different ways people understand, experience, and conceptualise a phenomenon (Larsson & Holmström , 2007; Marton , 1994) . It emphasises the individual as agent who interprets his/her own experiences and who actively creates an order to his/her own existence. It therefore facilitated the exploration of the ‘qualitatively different ways’ in which professional competence and associated CPD “are experienced, conceptualised, understood, perceived and apprehended” (Marton , 1994 , p. 4424) . ‘Bracketing’ is central to the phenomenographic approach and requires the researcher to effectively suspend research theories, previous research findings, researcher understandings, perceived notions, judgements, biases and own experience of a research topic (Merleau-Ponty , 1962) . This ensures “phenomena are revisited, freshly, naively, in a wide-open sense” (Moustakas , 1994 , p. 33) “in order to reveal engaged, lived experience” of research participants ( Merleau-Ponty , 1962 cited in Ashworth , 1999 , p. 708 ). In turn, participant experiences and understandings are examined and “characterised in terms of ‘categories of description’, logically related to each other, and forming hierarchies in relation to given criteria” (Marton , 1994 , p. 4424) . Such conceptions are assumed to have both meaning, a ‘what’ attribute, and structure, a ‘how’ attribute (Marton , 1994) . The anticipated output from Researcher B’s study sought an understanding of professional competence (the ‘what’ attribute) and the manner in which individual practitioners achieve and maintain such competence (the ‘how’ attribute). Interviews were considered the most appropriate data collection method to gain this understanding. The professional status of practitioners was therefore central to Researcher B’s study and the research focused on gaining an understanding of individual perceptions and practices with regard to maintaining and further developing professional competence. Mindful of this focus, the following research questions were developed:

What does it mean to be a ‘professional’?

What does ‘professional competence’ mean?

How is professional competence maintained and developed?

4. The NVivo and Manual Approaches to Qualitative Data Analysis

While Researchers A and B addressed disparate research areas, the above discussion indicates that qualitative data analysis represented a significant and central component of both researchers’ doctoral studies. Both researchers adopted an interpretivist philosophy involving a broadly similar number of interviews (27 in the case of Researcher A and 23 in the case of Researcher B). Despite the similarities between Researchers A and B, their choice of approach to qualitative data analysis was fundamentally different, with Researcher A choosing the technological approach (i.e. NVivo) and Researcher B the manual approach. In the remainder of this section, we discuss the factors influencing the choices made by Researchers A and B and provide insights into the data analysis process conducted. We then present critical reflections and the challenges faced by both researchers, as they undertook their respective approaches to qualitative data analysis.

4.1 Researcher A: Factors Influencing Approach to Qualitative Data Analysis

A number of factors influenced Researcher A’s decision to use NVivo (version 12) over the manual approach of qualitative data analysis. The most prominent of these was the multidimensional nature of the data collected. Researcher A investigated the financial decision-making process of farmers by exploring both strategic and operational decision-making. The farmers interviewed operated different farm types, had diverse levels of formal education and their age profile varied. The presence of multiple attributes highlighted the importance of reporting findings not only on how individual farmers undertook decision-making, but also to engage in comparisons of decision-making in different types of farming, and to explore how demographic factors (e.g. education, age) affected farmers’ decision-making processes.

Researcher A explored the option of adopting a technological approach to data analysis at an early stage in his study by attending a training course on NVivo. Despite attending the training course with an open mind and being aware of the alternative manual approach of qualitative data analysis, the training course convinced Researcher A of the potential power of NVivo to assist in qualitative data analysis. In particular, Researcher A was drawn to the ‘slice and dice’ capability of NVivo, whereby data could be analysed for a specific type of decision (strategic or operational), across multiple farm types (dairy, tillage or beef), or with respect to the demographic profile of farmers (education, age). By setting up different types of decisions, farm types and demographic factors as overarching themes (referred to as ‘nodes’ in NVivo), NVivo presented Researcher A with the ability to conduct numerous queries to address the research objectives, whilst simultaneously facilitating the extraction of relevant quotations to support findings. While the analysis could have been conducted manually, the search facility within NVivo was considered by Researcher A to be a very useful function and more efficient than using word processing software, which would be used with a manual approach. An additional and related factor which influenced Researcher A’s decision to proceed with NVivo was the possibility of availing of on-going one-to-one support for the duration of the research project from an NVivo trainer, when the actual qualitative data analysis commenced. In addition, Researcher A’s decision to opt for NVivo was influenced by his supervisor’s experience when conducting her own PhD studies. To that end, Researcher A’s supervisor had experience of using a technological approach (NUD*IST) to undertake qualitative data analysis. As a result of her familiarity with a technological approach, and an overall relatively positive experience, Researcher A’s supervisor provided some reassurance that this approach, versus the manual approach, was appropriate.

Before finally making the decision to adopt either a manual or technological approach to qualitative data analysis, Researcher A engaged with the various academic debates in the literature concerning the appropriateness of both. Based on these debates, Researcher A was confident that the technological approach to qualitative data analysis was appropriate. However, reflecting the debates in the literature, Researcher A was particularly mindful that “[NVivo] is merely a tool designed to assist analysis” (O’Dwyer , 2004 , p. 395) and that data analysis is ‘messy’ and very much the responsibility of the researcher who “must ask the questions, interpret the data, decide what to code” (Bringer et al. , 2006 , p. 248) .

4.2 Researcher A: An NVivo Approach to Data Analysis

Researcher A conducted 27 in-depth semi-structured interviews with farmers to develop an understanding of their financial decision-making processes. As with any qualitative research project, prior to formal data analysis, there was a significant amount of work involved in ‘cleansing’ the interview data collected. Researcher A transcribed all interview recordings, after which transcriptions were listened to and carefully read to identify inaccuracies. Field notes were also written by Researcher A immediately after each interview and these complemented the analysis of qualitative data and assisted the researcher in being reflexive during the data analysis process.

Researcher A adopted a thematic approach to qualitative data analysis as advocated by Braun & Clarke (2006) . Thematic analysis is a method for identifying, analysing and reporting patterns (themes) within data, where a theme is “something important about the data in relation to the research question and represents some level of patterned response or meaning from the data set” (Braun & Clarke , 2006 , p. 80) . In undertaking qualitative analysis, Researcher A followed a six phase thematic data analysis process (see Figure 1 ) developed by Braun & Clarke (2006) as follows:

Familiarising yourself with your data – interview transcripts were read and re-read by Researcher A, noting down initial ideas. Interview transcripts were then imported into the data management software NVivo.

Generating initial codes – this phase, descriptive coding, involved the deconstruction of the data from its initial chronology. The inductive process resulted in 227 hierarchical codes identified from the interview data, across 11 areas.

Searching for themes – this phase involved reviewing the open coding, merging, re-naming, distilling and collapsing the initial codes into broader categories of codes. This allowed the data to be constructed in a manner that enabled the objectives of the research to be fulfilled. Phase 3 resulted in the generation of 11 empirical themes related to strategic decision-making and 10 related to operational decision-making.

Reviewing themes – a process of ‘drilling down’ was conducted, including re-coding the text in the initial codes, re-organising into a coding framework, and breaking the themes down into sub-codes to better understand the meanings embedded therein.

Defining and naming themes – this involved abstraction of the data into a broader thematic framework. Using an inductive process, data was coded in relation to the four components of research objective 1, namely influencing factors; role of advisors; role of FFM; and other issues.

Producing the report – the final phase involved writing analytical memos to accurately summarise the content of each theme and propose empirical findings. The analytical memos helped Researcher A to produce a timely written interpretation of the findings, with the addition of his own annotations and recollections from interviews. The analytical memos also greatly assisted Researcher A to draft the findings chapter of his PhD thesis.

Figure 1

4.3 Researcher A: A Critical Reflection and Challenges with NVivo Qualitative Data Analysis

Reflecting on the journey of using NVivo as an approach to qualitative data analysis, Researcher A observed a number of salient points. First, a considerable amount of time and commitment is involved in developing the necessary skills to use the technology. Initially some time and effort are needed to learn how to operate the technology and formal NVivo training provides an essential support mechanism in this regard, particularly where training utilises standardised test data. Formal training also provides the researcher with an excellent overview of the technology and its potential capabilities. However, Researcher A cautions that it is not until the researcher actually begins to analyse their own data, which could potentially be some months/years later given the nature of the PhD research process, that specific study-related queries in using NVivo emerge. Due to the potential time lag, the researcher may have forgotten many aspects covered during the training or they may encounter queries that they have not experienced before. Hence, further specific guidance and/or further training may be required from the service provider. On a positive note, Researcher A found that the significant time and commitment invested towards the beginning of the data analysis process reaped considerable benefits towards the latter end of the research project. In particular, the systematic and structured coding process conducted allowed the retrieval of multi-layered analyses of the data relatively quickly. Furthermore, NVivo enabled the researcher to analyse, compare and contrast various aspects of the data efficiently and effectively. This was particularly useful for Researcher A, given the multidimensional aspect of the data collected. The time invested in learning how to operate the technology is a transferable research skill that the researcher could use on future research projects. While Researcher A invested a considerable amount of time becoming proficient with NVivo, it should be noted that the cost of both the technological approach (licence fee for NVivo) and formal training was not an issue, as these were funded by the researcher’s institution.

Second, critical reflection by Researcher A highlights the need to be mindful of the inclination to quantify qualitative data when using data analysis technologies. To that end, the coding process undertaken when using NVivo has the potential to focus the researcher’s attention on counting and quantifying the number of times a particular issue is identified or emphasised in the data. Braun & Clarke (2006) highlight that there are no hard and fast rules on how to identify a theme during qualitative data analysis. One cannot quantify how many times an issue must appear in the data in order for it to be labelled a theme. Indeed, an issue may appear infrequently in a data set, yet be labelled as a theme. Therefore, researcher judgement is necessary in determining themes. During the initial stages of writing up the findings, Researcher A found the above to be a particular challenge, as NVivo focused his attention on counting the number of times a particular issue appeared in the data. The ‘counting’ of data can be done easily through NVivo via the generation of graphs, tables or charts at the ‘push of a button’. Such analyses are useful for presenting a high-level overview of issues emphasised in the data, but they can also distract from the richness of the underlying interview data. Reflecting on this, Researcher A identified that it was necessary to pause, refocus and consider the underlying essence of the interview data, alongside the more quantitative output that NVivo generates. This is an important issue that qualitative researchers need to be cognisant of, particularly those who are first time users of the technological approach to analysing qualitative data.

Third, Researcher A reflects that the coding and analysis of the large volume of qualitative data collected was challenging and there was a need to be tolerant of uncertainty during this process. In particular, there was an element of drudgery and repetitiveness in coding the data using NVivo, necessitating the need for resilience and a ‘stick with it’ attitude as it was necessary to consistently code all interview data. However, one of the main benefits of adopting a systematic process, such as that facilitated by NVivo, is that it provides a map and audit trail of how the coding and analysis process was conducted. To some extent, this helped to structure the “messiness” (O’Dwyer , 2004 , p. 403) that is often attributed to qualitative data analysis.

Finally, reflecting on his overall experience, Researcher A found the NVivo data analysis software to be an excellent tool in terms of its ability to organise and manage qualitative data. In particular, the structured and systematic process of data analysis was very useful and effective. It is, however, important to note that while NVivo is a useful tool, it cannot replace the researcher’s own knowledge of the empirical data or the high level of research skills and judgement required to comprehend the data and elucidate themes, or the need for the researcher to be reflective in the data analysis process. In conclusion, Researcher A’s experience suggests that the benefits of using NVivo during the qualitative analysis phase outweigh the challenges it poses. Additionally, given the benefit of hindsight, Researcher A would use this technology in future qualitative research projects.

4.4 Researcher B: Factors Influencing Approach to Qualitative Data Analysis

A review of pertinent literature (Ashworth & Lucas , 2000; Larsson & Holmström , 2007; Svensson , 1997) highlights that there is no one ‘best’ method of phenomenographic data analysis. The overriding objective is to describe the data in the form of qualitative categories. This necessitates an approach for data analysis that enables resulting themes to be grounded in the data itself, rather than in prior literature or the researcher’s own experiences. However, Svensson (1997) cautions against replicating quantitative methodological traditions which view categories as “predefined assumptions” (p. 64). Mindful of this, and conscious that only a small number of phenomenographic studies had adopted a technological approach to data analysis at the time that Researcher B was making her decision on whether or not to adopt a technological approach (e.g. Ozkan , 2004) , Researcher B selected a non-technological manual approach. A further factor impacting on Researcher B’s decision to proceed with the manual approach was a perception that technological approaches, such as NVivo, were not used extensively by qualitative researchers within the Higher Education Institution in which she was enrolled as a PhD student. Whilst completing her doctorate studies at a UK University on a part-time basis, Researcher B attended a number of research methodology training sessions (funded by the researcher’s institution) and research seminars. Researchers who presented their work had adopted a manual approach to qualitative data analysis and were not very knowledgeable in relation to technological approaches. This highlighted an absence of an established community of practice in this regard and could mean that any adoption of a technological approach might not be appropriately aligned with the research community.

The experience of Researcher B’s supervisory team also influenced her decision to adopt the manual approach of qualitative data analysis. To that end, Researcher B’s supervisory team had no experience of using a qualitative technological approach for data analysis. This problem was compounded in that the supervisory team also had limited experience of qualitative research and was therefore reluctant to recommend any specific approach to data analysis. Taking on board the above factors, Researcher B believed there was no compelling reason to adopt a technological approach, thus she was not positively disposed towards NVivo or other such technological tool for qualitative data analysis. As a result, Researcher B selected a manual approach to qualitative data analysis.

4.5 Researcher B: A Manual Approach to Data Analysis

Researcher B was conscious of the “inevitable tension between being faithful to the data and at the same time creating, from the point of view of the researcher, a tidy construction useful for some further exploratory or educational purpose” (Bowden & Walsh , 2000 , p. 19) . Reflecting this, the analysis phase sought to gain insights into interview participants’ perceptions, meanings, understandings, experiences and interpretations. Consistent with the phenomenographic approach, Researcher B was mindful of the need for conscious bracketing with reference to the analysis of the interviews. [2] This comprised careful transcription of interviews, with emphasis on tone and emotions, and simultaneous continuous cycles of listening to interview recordings and reading of interview transcripts to highlight themes.

Researcher B found “the path from interviews through inference to categories…quite a challenge” (Entwistle , 1997 , p. 128) . The substantial volume of interview data required multiple and simultaneous continuous cycles of reading, note-making, interpretation, write-up and reflective review and the overall analysis of hard copy transcripts was quite a “messy” process (O’Dwyer , 2004 , p. 403) . It comprised substantial participant quotes highlighted in an array of colours on transcripts, a large amount of handwritten suggested thematic descriptions on both left and right transcript margins and large quantities of post-it notes of varying shades attached to the transcripts.

In undertaking the manual qualitative data analysis, Researcher B methodically worked through a series of steps, based on the work of Lucas (1998) and Ashworth & Lucas (2000) , as follows:

Familiarising self with the interviewee data and highlighting initial themes – Researcher B initially read each transcript a number of times and highlighted what she considered important elements of text with highlighter marker. She re-read each transcript a number of additional times and noted possible themes by writing on the right-hand margin of the hard copy transcript. She then highlighted more broad-based themes in the left-hand margin. Following this initial thematic identification, Researcher B re-read and listened to the interview recordings several more times, re-examining the analysis with a view to being methodical, yet open-minded about the content of the interviews.

Grounding themes in individual interviewee contexts – while many aspects of analysis focus on comparative experiences and mindful that these are of value, the phenomenographic approach positions individual experiences and lifeworlds as a backdrop to meanings. It was therefore important that individual experiences were not lost in an attempt to understand more generalising aspects. To this end, Researcher B also compiled individual interviewee profiles. The over-riding objective of this was to identify and examine particular points of emphasis that appeared to be central to the overall individual experiences with regard to development of professional competence. Such in-depth examination helped focus on the participants’ experiences and contributed to the empathetic understanding of participant perceptions, experiences, understandings and meanings (Lucas , 1998) . This also helped to counter tendencies to “attribute meaning out of context” (Lucas , 1998 , p. 138) and provided a means to understand participants’ experiences over a considerable period of time, from the point at which they made the conscious decision to gain admittance to the accounting profession up to the present day. This added considerable value to the analysis, not only helping to reveal what participants’ experiences and understandings of professional competence and professional development were, but also how participants shaped their ongoing actions and engagement with the development of professional competence. Predominant themes were then highlighted on the individual transcripts for each participant, in the participants’ own words. This served to maintain the bracketing process and ensured that themes were grounded in participants’ experiences.

Drafting initial thematic write-up – Researcher B drafted an initial descriptive thematic write-up, focussed around the research questions.

Reviewing interview data for supporting quotes – relevant interviewee quotes for each theme were subsequently included in the draft thematic write-up.

Reviewing thematic write-up – Researcher B re-read and listened back to the interviews several more times. She also searched individual interview transcript word documents for key words and phrases to highlight additional quotes to support thematic descriptions. She then spent some time editing the write-up with a view to generating a more “tidy construction” of descriptive overall categories (Bowden & Walsh , 2000 , p. 19) .

Generating categories of description – the final stage of analysis was the generation of overriding categories of description . The what aspect was used to characterise what professional competence means to participants (i.e. the meaning attribute) while the how aspect categorised how participant practitioners actually maintain and develop their professional competence (i.e. the structural attribute). Participants’ experiential stages were used to inform the hierarchy vis-a-vis these categories.

4.6 Researcher B: A Critical Reflection and Challenges with Manual Qualitative Data Analysis

Researcher B reflects on the challenges pertaining to data analysis during the course of her PhD study and highlights a number of issues. While the manual approach facilitated the generation and analysis of themes from the interview data, it was challenging to manage themes that were continuously being defined and redefined. Notwithstanding the iterative nature of the manual approach, Researcher B was confident that themes developed in an organic manner and were not finalised too early in the data analysis process. The ambiguity associated with the generation and analysis of themes also required Researcher B to bring high levels of research knowledge and skills to support this process and to be mindful of the need to embrace high levels of tolerance for uncertainty. Researcher B acknowledges that the iterative process of reading interviewee transcripts, listening to interview recordings (largely while in the car on the commute to and from work or while taking trips to see family at the other side of the country), generating themes, writing up themes, followed by re-reading messy transcripts and re-listening to the interview recordings while re-visiting themes, was both tedious and time consuming.

The initial excitement experienced when first listening to the interview recordings and reading the interview transcripts was somewhat depleted by the end of the process and work on the analyses increasingly developed into a test of endurance. Researcher B likened this to the declining enthusiasm often experienced by students from first reading a clean copy of a Shakespearian play in school, followed by subsequent grappling with syllabus requirements to dissect the play in multiple different ways in order to isolate significant events, explore characters, interpret language, examine subplots and understand larger themes. At the end of the school year, the once clean hard copy has become a heavily annotated and much more complex version of the original and the students’ enthusiasm considerably more subdued.

Researcher B also reflects that the manual approach required her to become very familiar with the interviewee transcripts and recordings, such that Researcher B could effectively match interview quotes to interviewees without having to check their provenance. Researcher B acknowledges that some participants provided more considered and more articulate responses to interview questions, and on review of the initial draft write-up, realised she had included excessive quotes centred around such participants. In subsequent iterations, Researcher B was careful to ensure the write-up was more representative of all of the interviewees and not dominated by a small number of interviewees.

As analysis progressed during the course of the doctorate, Researcher B presented draft write-ups of her findings to her PhD supervisors at various stages, largely to seek reassurance that data analysis was progressing appropriately. However, as indicated earlier, both supervisors had limited experience of qualitative data analysis and could provide little categorical reassurance regarding the manual approach to data analysis. As such, Researcher B had no systematic source of affirmation and was prompted to present at various doctoral colloquia to gain further insights and validation of the approach to analysis. This provided a useful, albeit more ad hoc , source of guidance and affirmation.

Finally, Researcher B reflects on the overall doctoral process and more particularly on the selection of a manual approach to data analysis. With hindsight, she recognises that while this approach enabled closeness to the interview data, data management involved a significant amount of time. For example, ‘cutting’ and ‘pasting’ within word documents which had to be done and re-done many times, reflecting the messiness of the data analysis. This was quite repetitive and was not an efficient means of organising data to support research findings. Researcher B believes that qualitative data analysis should enable both a closeness to the data and an efficient means of managing data. To that end, she would consider trialling measures to enhance the efficiency of data management in future research studies, including use of software tools such as NVivo.

5. Discussion and Conclusion

This paper addresses a gap in the literature by providing reflective and critical insights into the experiences of two PhD researchers undertaking qualitative studies which adopted different approaches to data analysis. The experiences and reflections of Researchers A and B highlight some similarities and differences worthy of note. In terms of background and motivations, while both researchers were investigating different research areas, qualitative data analysis was a central and shared aspect of both. To that end, both researchers were faced with the same decision regarding the choice of qualitative data analysis approach, Researcher A deciding on a technological approach (NVivo) and Researcher B opting for the manual approach.

Table 1 summarises the factors influencing the choice of data analysis approach adopted by Researchers A and B, together with the challenges and benefits of each. Interestingly, while the similarities in background and motivations detailed in the paper had little impact on both researchers’ decision regarding the qualitative data analysis approach, the factors influencing the choice were markedly different. To that end, Researcher B’s engagement with a more extensive literature exploring phenomenographic data analysis indicated that few prior studies had adopted a technological approach. Coupled with the lack of a community of practice with experience of using the technological approach, these factors were primary influences on Researcher B’s decision to adopt a manual approach. This decision has some parallels with O’Dwyer’s (2004) experience of discounting the technological approach at an early stage of his research based on his lack of understanding of what it could offer. In contrast, Researcher A’s decision-making process was largely influenced by the multi-dimensional nature of the interview data collected and exposure to an NVivo training course where the potential of the software’s ‘slice and dice’ and query capabilities were demonstrated. The possibility of accessing on-going NVivo one-to-one support for the duration of the research project was a further factor in Researcher A’s decision to use the technological approach. While different factors clearly influenced Researchers A and B’s decision regarding their qualitative data analysis approach, the experiences of their supervisory teams were common to both. Researcher A was influenced to ado,pt the technological approach as a result of his supervisor’s positive experience, while Researcher B was influenced to adopt the manual approach due to her supervisors’ limited knowledge or experience of the technological approach. This finding points to the importance of supervisors’ experience in informing the decision regarding the qualitative data analysis approach and highlights a potential danger of narrowing the data analysis choices available to the doctoral researcher.

The critical reflections of both researchers also elucidate some key challenges and benefits that qualitative researchers should be mindful of. Despite adopting different approaches, both researchers highlighted challenges in terms of the time consuming and labour intensive nature of their respective data analysis approaches, largely consistent with earlier findings (Bédard & Gendron , 2004) . While Researcher A had to invest considerable time and commitment in developing the skills required to use NVivo, this reaped significant benefits towards the latter end of his research project in terms of the efficient retrieval of information, confirming previous literature (Basit , 2003; García-Horta & Guerra-Ramos , 2009; Zamawe , 2015) . Researcher B also noted a challenge around the time-consuming nature of the data analysis process using the manual approach and the significant investment in time for activities such as listening to recordings, reading and re-reading of transcripts, and ‘cutting’ and ‘pasting’ which had to be done and re-done, again consistent with earlier research findings (Basit , 2003; Bogdan & Bilken , 1982; Lofland , 1971; Maher et al. , 2018; L. Richards & Richards , 1994) . Researcher A’s experience, however, highlights a further challenge not identified in the prior literature with respect to investment in time, namely the resulting time lag that can occur between the timing of initial NVivo training and the actual use of the technology, with the result that important knowledge and skills relevant to analysis have been ‘forgotten’.

Both researchers also highlighted an element of drudgery and repetitiveness in coding their data and developing themes, and the need for resilience (Researcher A) and endurance (Researcher B) in this regard. Drawing on their experiences, both researchers were mindful of “being tolerant of uncertainty [which] is part of the fundamental skills of the qualitative researcher” (Bédard & Gendron , 2004 , p. 199) . Irrespective of the approach to qualitative data analysis, both Researchers A and B were also cognisant of the importance of retaining a level of ‘closeness’ to their data and an awareness that the approach to analysis cannot substitute for the researcher’s own knowledge of the empirical data (O’Dwyer , 2004) . Furthermore, Researchers A and B’s experiences provide new insights to the literature. Researcher A recognised the potential danger of NVivo over-focusing the researcher’s attention on counting and quantifying and how this might negatively impact in terms of maintaining a level of closeness with the data. In addition, Researcher B cautioned against the possibility of being ‘too close’ to some interviewee data when using a manual approach, and the need to continually and consciously ensure that the qualitative data analysis was representative of all interviewees. Reflecting further on the tedious nature of the manual process, Researcher B reported an additional challenge in that a significant amount of time had to be devoted to data management activities (i.e. cutting and pasting into word documents) given the ‘messiness’ of her data analysis.

Both researchers identified some benefits of their respective data analysis approaches. Researcher A recognised that the technological approach, NVivo, provides a systematic coding process with a clear audit trail which helps to structure the ‘messiness’ attributed to qualitative data analysis (O’Dwyer , 2004 , p. 403) . In addition, Researcher A highlighted that NVivo is an excellent tool in terms of its ability to organise and manage qualitative data. The skills developed as a result yield significant benefits in terms of facilitating multi-layered analyses that can be used in future research projects. In contrast, Researcher B reflected on how the manual approach facilitated a closeness to the qualitative data (notwithstanding the challenge highlighted earlier in this regard) and that this approach facilitated the identification of themes in an organic manner.

The preceding discussion lends support to the conclusion that the choice of a manual or technological approach to qualitative data analysis is influenced by multiple factors. In making a decision regarding the approach to data analysis, researchers need to be cognisant of the potential challenges and benefits of their choices. Ultimately, however, the final decision regarding the approach to adopt is a personal choice. Irrespective of the choices available to the researcher, it is important to acknowledge that qualitative data analysis is “the most intellectually challenging phase” of qualitative research (Marshall & Rossman , 1995 , p. 114) . Described as ‘messy’ by O’Dwyer (2004) , qualitative data analysis is also labour intensive, requiring high levels of research knowledge and skills, and associated with the need to be tolerant of uncertainty (Bédard & Gendron , 2004) . The experiences and reflections of both researchers in this paper provide evidence of these challenges. While this paper provides insights into the choice of qualitative data analysis approach, a limitation is that it does not address how manual or technological approaches to qualitative data analysis consider issues related to the quality of data analysis undertaken. For example, Pratt et al. (2019) highlight the need to identify solutions for enhanced trustworthiness (an aspect of quality) in qualitative research. Further research might consider how the manual and technological approaches address such issues. Another limitation of the paper is that the experiences outlined reflect those of two individual researchers. These experiences may not be reflective of the experiences of others who engage in the manual or technological approaches to qualitative data analysis. Further research which more broadly compares the experiences of other qualitative researchers would add greater insights in this under-researched area.

The paper contributes to the limited literature on the comparative experiences of those undertaking qualitative data analysis using the manual and technological approaches. In so doing, we identify the factors influencing the choice of approach, confirming in some respects prior findings in the literature, but also adding to the small body of prior literature. We further contribute to the limited literature by adding insights into the challenges and benefits of the manual and technological approaches to qualitative data analysis. “Given the popularity of interviews as a method of qualitative data collection in accounting” (Lee & Humphrey , 2006 , p. 188) , the paper adds insights into how researchers address one of the key problems they face, namely how to analyse interview transcripts using the manual and technological approach. We thereby respond to calls from Edwards & Skinner (2009) and Paulus et al. (2017) for future studies to provide insights into qualitative researchers’ experiences of using the manual and technological approaches to data analysis. We hope that the experiences and reflections articulated in this paper, including the factors impacting on and the challenges and benefits of using the manual and technological approaches, will guide qualitative researchers in making important decisions regarding their approach to data analysis. The issue of how to analyse qualitative data, and whether to use manual or technological approaches is often a source of difficulty for researchers, we hope that this paper will initiate further debate around this important decision.

The manual approach involves analysing qualitative data without the use of computerised data analysis software.

The issue of bracketing is a core element of the phenomenographic research approach, irrespective of the selection of a manual or a technological approach to data analysis.

An official website of the United States government

Official websites use .gov A .gov website belongs to an official government organization in the United States.

Secure .gov websites use HTTPS A lock ( Lock Locked padlock icon ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.

  • Publications
  • Account settings
  • Advanced Search
  • Journal List

Sage Choice logo

Secondary Data Analysis in Nursing Research: A Contemporary Discussion

Siobhan o’connor , lecturer, bsc, cima cba, bsc, rn, fhea, phd.

  • Author information
  • Article notes
  • Copyright and License information

Dr Siobhan O’Connor, Lecturer, BSc, CIMA CBA, BSc, RN, FHEA, PhD, School of Health in Social Science, The University of Edinburgh, Doorway 6 Old Medical Quad, Teviot Place, Edinburgh EH8 9AG, UK. Email: [email protected]

Issue date 2020 Jun.

This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License ( https://creativecommons.org/licenses/by-nc/4.0/ ) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access page ( https://us.sagepub.com/en-us/nam/open-access-at-sage ).

This editorial provides an overview of secondary data analysis in nursing science and its application in a range of contemporary research. The practice of undertaking secondary analysis of qualitative and quantitative data is also discussed, along with the benefits, risks and limitations of this analytical method.

Keywords: secondary analysis, analysis, nursing, research

Introduction

The earliest reference to the use of secondary data analysis in the nursing literature can be found as far back as the 1980’s, when Polit & Hungler (1983 ), in the second edition of their classic nursing research methods textbook, discussed this emerging approach to analysis. At that time, this method was rarely used by nursing researchers. McArt & McDougal (1985 ) posit a number of reasons for the lack of secondary data analysis in nursing at that point including a preference for empirical research, limited datasets available in healthcare making it less favourable, and low awareness or appreciation of this type of analysis. Its origins lay further back within the wider educational, social science, and other scientific literature as Glass (1976 ) described the process as ‘ re-analysis of data for the purpose of answering the original research question with better statistical techniques, or answering new questions with old data ’. Over 40 years later, it seems nursing science has come full circle with secondary data analysis widely employed across many areas of clinical, pedagogical, and policy research ( Aktan, 2012 ; Naef et al., 2017 ). So, what has changed?

One paradigm shift over the preceding decades has been the digitisation of data, driven by advances in computing. Early forms of modern day Electronic Health Records (EHRs) and other hospital information systems emerged in the United States in the 1960’s and 1970’s, and their use slowly spread worldwide ( Musen & van Bemmel, 1997 ). These enable researchers to tap into a wealth of clinical and administrative hospital data for secondary analysis. For example, nurses have examined electronic care plans to determine if they meet national documentation standards ( Häyrinen et al., 2010 ) and used EHRs in home health agencies to identify interventions that could improve urinary and bowel incontinence ( Westra et al., 2011 ). As the World Wide Web, more commonly known as the Internet, became accessible to the public in the 1990’s, researchers were able to utilise this new global, communications tool to share and access health datasets more easily. In tandem, online environments themselves, particularly social media platforms, created new virtual forums where patients, carers, health professionals and others could interact. This permits researchers to mine these new data sources in numerous ways. For instance, nurses have investigated patient and family blogs about illness to enhance online communication ( Heilferty, 2009 ) and used Twitter datasets to appreciate how social media could be employed to inform health policy ( O'Connor, 2017 ).

The rapid developments in computing along with those in telecommunications led to the rise of mobile technology in the 1990’s and 2000’s. Smartphones and accompanying health applications give patients and the public the ability to collect their own personal health data for self-management and self-care ( Heidi et al., 2017 ). Mobile devices and applications are also used by health professionals and students for a range of purposes such as monitoring patient vital signs, prescribing medication, and clinical decision making ( O’Connor & Andrews, 2018 ; Ventola, 2014 ). This allows researchers to employ citizen science techniques to gather data from apps for secondary analysis. To illustrate, nurses developed analytics for an app to track tobacco use in psychiatric patients ( Oliveira et al., 2016 ) and monitored how health professionals engage with apps through Google and other analytics platforms ( Maskey et al., 2013 ). Wearable and sensor technologies for personal and home health monitoring were next to follow, adding to the health datasets potentially available for re-analysis ( Sloan et al., 2018 ). This is not to say that paper based forms of information such as patient diaries are not valuable sources and are still being used for secondary analysis ( Cheraghi-Sohi et al., 2013 ).

Making the most of digital data by linking datasets to enable ‘Big Data’ analysis ( O'Connor, 2018 ) is now considered by some to be the epitome of secondary data analysis in many areas of science. Big Data is often described using five ‘Vs’; volume, velocity, variety, veracity and value, reflecting the types of datasets it can encompass and the challenges of analysing these. This approach is being utilised in nursing in many ways such as mining EHRs, web-based reporting systems, and clinical and organisational databases by employing a range of statistical and other techniques ( Westra et al., 2017 ). Now, we also have more sophisticated software tools such as Hadoop and Tableau (see Figure 1 ) that enable secondary analysis of digital datasets or some researchers use programming languages such as Python or R to create their own specialised analytical tools ( Bogdan & Raluca Mariana, 2014 ). Furthermore, visualisation is becoming a popular approach to presenting the results of this process. Although data visualisation was pioneered by Florence Nightingale over 150 years ago ( O'Connor et al., 2020 ), it is now being widely applied to augment primary and secondary analysis so that complex findings can be presented in clear and coherent ways.

Figure 1.

Screenshot of Tableau software with a data visualisation.

Source/Credit: By Marissa-anna - Own work, CC BY-SA 4.0, https://commons.wikimedia.org/w/index.php?curid=74539538 .

Outside of the major technological shifts and digitisation of data in recent years, secondary data analysis has also become more widely used due to the challenges of undertaking empirical research. An issue some nursing scientists face is recruiting populations of patients or carers that are difficult to reach due to a myriad of social, cultural, economic and political reasons. These may include refugee and migrant groups, those who experience domestic and sexual violence, homelessness and many more ( Biederman & Forlan, 2016 ). Research fatigue in over-researched groups that may include some cancer patients, certain indigenous communities, nursing students, and others can also be avoided by adopting secondary analysis ( Clark, 2008 ). Hence, utilising existing datasets related to participants of interest can offer an alternative way to examine some issues while removing respondent burden ( Ziebland & Hunt, 2014 ). Szabo & Strang (1997 ) suggest it can also help reduce researcher bias and provide some objectivity, as the researcher may not have been immersed in the original study design or data collection. Equally, accessing different professional groups from clinicians to policy makers may prove difficult at times. For instance, emerging global health crises such as COVID-19 pose barriers to recruiting these types of participants and carrying out primary data collection, outside of research focused on addressing the immediate health crisis ( Nicol et al., 2020 ). Therefore, tapping into existing datasets and interpreting them to address research questions can be beneficial, enabling an area of nursing science to move forward.

Another important factor driving more secondary analysis of data is the emergence of open data initiatives and policies that promote open access to scientific data and research. The open data movement has been gathering pace for several years as governments make public datasets more easily accessible to accelerate data driven innovation and gain greater returns on the funds invested in research ( Conway & Vanlare, 2010 ). For example, the Human Genome Project was a major international research initiative in the 1990’s and early 2000’s that modelled the full human genome. It made this data freely available to facilitate knowledge and scientific discovery, leading to new fields such as personalised medicine ( Collins & McKusick, 2001 ) which has many implications for nursing ( Vorderstrasse et al., 2014 ). Open access is now encouraged by many research funders so that datasets collected are placed in a freely available online repository for others to use. This culture of data sharing is likely to continue given the large amounts of funding required for empirical research as secondary analysis can be a cost-effective way to uncover new insights, particularly when there is a challenging financial landscape to contend with.

The other driver of secondary data analysis in nursing can be attributed to developments in the professionalisation of nursing education, clinical practice and research. With a move to graduate nursing education in many countries, evidence based practice is now taught to nursing students so they can incorporate the results of scientific research into future decision making and care delivery ( Mackey & Bassendowski, 2017 ). Masters and doctoral level study are now more widely available to nurses globally, with the volume of nursing PhD and Doctor of Nursing Practice programmes increasing in some regions ( Bednash, et al., 2014 ). Nowadays, nurses often have extra opportunities to undertake research throughout their clinical and academic careers ( Currey et al., 2011 ; Francis & Humphreys, 1999 ). This means that more primary data in nursing and healthcare is being gathered that can form the basis for secondary analysis. Furthermore, with added opportunities for advanced training, nursing science is broadening its traditional base and expanding its ontological, epistemology and methodological expertise facilitating more secondary data analysis.

The Practice of Secondary Data Analysis

Given the widespread and growing use of secondary data analysis in scientific research, it is worthwhile revisiting this practice to appreciate how it can generate evidence in nursing, now and into the future. Beck (2019 ) discusses some of the practical aspects to consider before beginning such as identifying appropriate datasets and negotiating access to these as this can take time and money. Assessing the quality of the secondary dataset is also recommended so its strengths and limitations can be understood, as this may impact the analytical methods used along with the findings and will need to be reported in any published works. This could include reviewing the expertise and qualifications of those involved in gathering and processing the data, considering any contextual information such as accompanying field notes or ethical approval, and examining the completeness of the primary data whether that is qualitative, quantitative or a mixture of both. Heaton (1998 ) also suggests considering the size and diversity of the sample to gauge whether it is adequate to address certain research questions, along with the availability of the original researchers to consult and provide guidance and clarification where needed.

In terms of the secondary analysis of qualitative data, Heaton (2004 ) describes five different ways to undertake this. Firstly, supplementary analysis is an in-depth analysis of an emergent concept in a qualitative dataset not fully explored in the primary study. For example, Hurlock-Chorostecki et al. (2013 ) used this approach on focus groups to uncover nurse practitioner interprofessional practice. While this retroactive interpretation can provide useful insights quickly and easily, it may be limited if the underlying dataset is not rich enough. Secondly, supra analysis uses existing qualitative data to address a new research question in a separate study. Sanna-Maria et al. (2017 ) employed this technique to examine the perceptions of nurse leaders on ethical recruitment in clinical research. Although this can allow for new perspectives and settings to broaden our understanding of a phenomenon, bias could be introduced if there is not a good ‘fit’ between the secondary data and the new research questions or study design ( Hinds et al., 1997 ). Thirdly, re-analysis relies on additional analyses of qualitative data to authenticate the results of a primary study. Saal et al. (2018 ) applied this in a mixed methods study to develop a complex intervention to improve social participation in care home residents with joint contractures. Even though re-analysis can strengthen the results of a primary study simply and speedily, the reinterpretation of qualitative data may lead to misconceptions and different findings ( Swanson, 1986 ).

Fourthly, amplified analysis occurs when two or more qualitative datasets are combined and then compared and contrasted using secondary analysis. Stickley et al. (2018 ) utilised this to examine the relationship between participatory arts and how people recover from mental health conditions. Granted this may provide richer datasets to investigate a phenomenon but by collectively pooling qualitative data some contextual or conceptual insights may be lost ( Sandelowski, 1991 ). Lastly, assorted analysis involves secondary analysis that is undertaken alongside the analysis of primary qualitative data. Watters et al. (2018 ) exploited this method to explore resilience and social inclusion among single mothers across Canada. Although analysing qualitative datasets in tandem could enrich and augment the final results, there is a risk of cross contamination during coding and analysis that might lead to inaccurate findings ( Heaton, 2004 ).

In terms of the secondary analysis of quantitative data, traditional approaches utilising descriptive and inferential statistics on an array of datasets are common. Secondary sources of quantitative data may include national census conducted by government, local or regional datasets held by public bodies, or questionnaires and surveys undertaken by researchers at a university or other type of national or international institution ( Dale et al., (2008 ). For example, Oh et al. (2016 ) mined the Korea Youth Risk Behaviour Web-based Survey to determine whether satisfaction with sleep was linked to stress in adolescents with atopic disease, while Jacoby et al. (2017 ) reused data from a longitudinal cohort study of psychological outcomes from minor injury to examine how this relates to recovery and disability. Digital archives held by libraries, museums or other social and cultural agencies could also be useful sources of quantitative data, with some providing an extensive catalogue that is searchable online. The International Federation of Data Organisations ( IFDO, 2020 ) and the Consortium of European Social Science Data Archives ( CESSDA, 2020 ) may be helpful in identifying national archives for secondary analysis. However, Dale et al. (2008 ) warn of potential problems with the secondary analysis of quantitative data as surveys and other measurement tools may have been constructed and their reliability and validity determined in specific ways. Equally the sample of participants, their characteristics and response rates may pose issues when modelling for correlation or causation. Hence, a critical eye should be cast to appreciate the strengths, limitations and biases inherent in a quantitative dataset before reusing it.

The power of combining qualitative and quantitative datasets in a mixed methods study design is becoming more evident within nursing science ( Hall et al., 2018 ), providing a richer ground within which to employ secondary data analysis. Dugas et al. (2017 ) adopted a sequential explanatory design and examined the results of a systematic review on the processes of developing patient decision aids, followed by interviews with vulnerable patients to better understand how to involve them in this type of research in the future. Newer statistical techniques such as machine learning, in particular deep learning, from the growing artificial intelligence community are also being developed and offer new ways to interpret secondary quantitative datasets. Although this is still a relatively novel approach in nursing, Bose & Radhakrishnan (2018 ) employed a variety of clustering techniques to model the characteristics of heart failure patients who used telehealth services from two home health agency datasets.

Secondary data analysis is now firmly embedded in nursing science, helping researchers to uncover new insights that can improve nursing education, patient care, health service delivery, public health, and health policy. No doubt new ways of approaching and conducting this analytical method will continue to emerge and become adopted into the practice of nursing research worldwide.

Acknowledgments

Author biography.

Dr Siobhan O’Connor , BSc, CIMA CBA, BSc, RN, FHEA, PhD, is a Lecturer in Nursing Studies at the University of Edinburgh, United Kingdom. She is a core member of the faculty, with a multidisciplinary background in both nursing and information systems. Hence, her research interests focus on the design, implementation, and use of technology in healthcare. https://www.research.ed.ac.uk/portal/en/persons/siobhan-oconnor(283996e9-3744-46c4-b4a2-dc2700e17297).html

Author Contributions: The sole author drafted and wrote the manuscript.

Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Funding: The author(s) received no financial support for the research, authorship, and/or publication of this article.

Inline graphic

  • Aktan N. M. (2012). Social support and anxiety in pregnant and postpartum women: A secondary analysis. Clinical Nursing Research, 21(2), 183–194. 10.1177/1054773811426350 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Beck C. T. (2019). Secondary qualitative data analysis in the health and social sciences. Routledge. [ Google Scholar ]
  • Bednash G., Breslin E. T., Kirschling J. M., Rosseter R. J. (2014). PhD or DNP: Planning for doctoral nursing education. Nursing Science Quarterly, 27(4), 296–301. [ DOI ] [ PubMed ] [ Google Scholar ]
  • Biederman D. J., Forlan N. (2016). Desired destinations of homeless women: Realizing aspirations within the context of homelessness. Creative Nursing, 22(3), 196 10.1891/1078-4535.22.3.196 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Bogdan O., Raluca Mariana D. (2014). Integrating R and Hadoop for big data analysis. Revista Română de Statistică, 62(2), 83–94. [ Google Scholar ]
  • Bose E., Radhakrishnan K. (2018). Using unsupervised machine learning to identify subgroups among home health patients with heart failure using telehealth. CIN: Computers, Informatics, Nursing, 36(5), 242–248. 10.1097/CIN.0000000000000423 [ DOI ] [ PubMed ] [ Google Scholar ]
  • CESSDA. (2020). Consortium of European Social science data archives. https://www.cessda.eu/
  • Cheraghi-Sohi S., Bower P., Kennedy A., Morden A., Rogers A., Richardson J., Sanders T., Stevenson F., Ong B. N. (2013). Patient priorities in osteoarthritis and comorbid conditions: A secondary analysis of qualitative data. Arthritis Care & Research, 65(6), 920–927. [ DOI ] [ PubMed ] [ Google Scholar ]
  • Clark T. (2008). ‘We're Over-Researched Here!’: Exploring accounts of research fatigue within qualitative research engagements. Sociology, 42(5), 953–970. 10.1177/0038038508094573 [ DOI ] [ Google Scholar ]
  • Collins F. S., McKusick V. A. (2001). Implications of the human genome project for medical science. JAMA, 285(5), 540–544. 10.1001/jama.285.5.540 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Conway P. H., Vanlare J. M. (2010). Improving access to health care data: The open government strategy. JAMA, 304(9), 1007–1008. 10.1001/jama.2010.1249 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Currey J., Considine J., Khaw D. (2011). Clinical nurse research consultant: A clinical and academic role to advance practice and the discipline of nursing. Journal of Advanced Nursing, 67(10), 2275–2283. 10.1111/j.1365-2648.2011.05687.x [ DOI ] [ PubMed ] [ Google Scholar ]
  • Dale A., Wathan J., Wiggins V. (2008). Secondary analysis of quantitative data sources. In Alasuutari P., Bickman L., Brannen J. (Eds.), The Sage handbook of social research methods. Sage. [ Google Scholar ]
  • Dugas M., Trottier M.-A., Chipenda Dansokho S., Vaisson G., Provencher T., Dogba M. J., Dupéré S., Fagerlin A., Giguere A. M., Haslett L., Hoffman A. S., Ivers N. M., Légaré F., Légaré J., Levin C. A., Menear M., Renaud J. S., Stacey D., Volk R. J., Witteman H. O. (2017). Involving members of vulnerable populations in the development of patient decision aids: A mixed methods sequential explanatory study.(Report). BMC Medical Informatics and Decision Making, 17(1). 10.1186/s12911-016-0399-8 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Francis B., Humphreys J. (1999). Enrolled nurses and the professionalisation of nursing: A comparison of nurse education and skill-mix in Australia and the UK. International Journal of Nursing Studies, 36(2), 127–135. 10.1016/S0020-7489(99)00006-1 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Glass G. V. (1976). Primary, secondary, and meta-analysis of research. Educational Researcher, 5(10), 3–8. [ Google Scholar ]
  • Hall H., Brosnan C., Cant R., Collins M., Leach M. (2018). Nurses’ attitudes and behaviour towards patients’ use of complementary therapies: A mixed methods study. Journal of Advanced Nursing, 74(7), 1649–1658. 10.1111/jan.13554 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Häyrinen K., Lammintakanen J., Saranto K. (2010). Evaluation of electronic nursing documentation—Nursing process model and standardized terminologies as keys to visible and transparent nursing. International Journal of Medical Informatics, 79(8), 554–564. 10.1016/j.ijmedinf.2010.05.002 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Heaton J. (1998). Secondary analysis of qualitative data. Social Research Update Issue 22. http://sru.soc.surrey.ac.uk/SRU22.html
  • Heaton J. (2004). Reworking Qualitative Data. SAGE Publications Ltd. [ Google Scholar ]
  • Heidi H., Wahl A. K., Småstuen M. C., Ribu L. (2017). Tailored communication within mobile apps for diabetes self-management: A systematic review. Journal of Medical Internet Research, 19(6). 10.2196/jmir.7045 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Heilferty C. M. (2009). Toward a theory of online communication in illness: Concept analysis of illness blogs. Journal of Advanced Nursing, 65(7), 1539–1547. 10.1111/j.1365-2648.2009.04996.x [ DOI ] [ PubMed ] [ Google Scholar ]
  • Hinds P., Vogel R., Clarke-Steffen L. (1997). The possibilities and pitfalls of doing a secondary analysis of a qualitative data set. Qualitative Health Research, 7(3), 408–424. [ Google Scholar ]
  • Hurlock-Chorostecki C., Forchuk C., Orchard C., Reeves S., van Soeren M. (2013). The value of the hospital-based nurse practitioner role: Development of a team perspective framework. Journal of Interprofessional Care, 27(6), 501–508. 10.3109/13561820.2013.796915 [ DOI ] [ PubMed ] [ Google Scholar ]
  • IFDO. (2020). International Federation of Data Organizations. http://www.ifdo.org/
  • Jacoby S. F., Shults J., Richmond T. S. (2017). The effect of early psychological symptom severity on long-term functional recovery: A secondary analysis of data from a cohort study of minor injury patients. International Journal of Nursing Studies, 65, 54–61. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Mackey A., Bassendowski S. (2017). The history of evidence-based practice in nursing education and practice. Journal of Professional Nursing, 33(1), 51–55. 10.1016/j.profnurs.2016.05.009 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Maskey M. M., Alexander M. S., Conover M. H., Gamble M. J., Hoy M. H., Fraley M. A. (2013). Powering transplant professional collaborations with web and mobile apps. CIN: Computers, Informatics, Nursing, 31(8), 351–355. 10.1097/CIN.0000000000000001 [ DOI ] [ PubMed ] [ Google Scholar ]
  • McArt E. W., McDougal L. W. (1985). Secondary data analysis–a new approach to nursing research. Image: The Journal of Nursing Scholarship, 17(2), 54–57. 10.1111/j.1547-5069.1985.tb01418.x [ DOI ] [ PubMed ] [ Google Scholar ]
  • Musen M. A., van Bemmel J. H. (1997). Handbook of medical informatics. Bohn Stafleu Van Loghum. [ Google Scholar ]
  • Naef R., Hediger H., Imhof L., Mahrer-Imhof R. (2017). Variances in family carers' quality of life based on selected relationship and caregiving indicators: A quantitative secondary analysis. International Journal of Older People Nursing, 12(2), n/a-n/a. 10.1111/opn.12138 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Nicol G. E., Piccirillo J. F., Mulsant B. H., Lenze E. J. (2020). Action at a distance: Geriatric research during a pandemic. Journal of the American Geriatrics Society. 10.1111/jgs.16443 [ DOI ] [ PMC free article ] [ PubMed ]
  • O’Connor S., Andrews T. (2018). Smartphones and mobile applications (apps) in clinical nursing education: A student perspective. Nurse Education Today, 69, 172–178. 10.1016/j.nedt.2018.07.013 [ DOI ] [ PubMed ] [ Google Scholar ]
  • O'Connor S. (2017). Using social media to engage nurses in health policy development. Journal of Nursing Management, 25(8), 632–639. 10.1111/jonm.12501 [ DOI ] [ PubMed ] [ Google Scholar ]
  • O'Connor S. (2018). Big data and data science in health care: What nurses and midwives need to know. Journal of Clinical Nursing, 27(15–16), 2921–2922. 10.1111/jocn.14164 [ DOI ] [ PubMed ] [ Google Scholar ]
  • O'Connor S., Waite M., Duce D., O'Donnell A., Ronquillo C. (2020). Data visualization in healthcare: The Florence effect. Journal of Advanced Nursing. 10.1111/jan.14334 [ DOI ] [ PubMed ]
  • Oh W. O., Im Y., Suk M. H. (2016). The mediating effect of sleep satisfaction on the relationship between stress and perceived health of adolescents suffering atopic disease: Secondary analysis of data from the 2013 9th Korea Youth Risk Behavior Web-based Survey. International Journal of Nursing Studies, 63, 132–138. 10.1016/j.ijnurstu.2016.08.012 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Oliveira R. M. d., Duarte A. F., Alves D., Furegato A. R. F. (2016). Development of the TabacoQuest app for computerization of data collection on smoking in psychiatric nursing. Revista Latino-Americana de Enfermagem, 24(0). 10.1590/1518-8345.0661.2726 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Polit D., Hungler B. P. (1983). Nursing research: Principles and methods (2nd ed.). Philadelphia: Lippincott. [ Google Scholar ]
  • Saal S., Meyer G., Beutner K., Klingshirn H., Strobl R., Grill E., Mann E., Köpke S., Bleijlevens M. H. C., Bartoszek G., Stephan A. J., Hirt J., Muller M. (2018). Development of a complex intervention to improve participation of nursing home residents with joint contractures: A mixed-method study.(Report). BMC Geriatrics, 18(1). 10.1186/s12877-018-0745-z [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Sandelowski M. (1991). Telling stories: Narrative approaches in qualitative research. Image:The Journal of Nursing Scholarship, 23(3), 161. [ DOI ] [ PubMed ] [ Google Scholar ]
  • Sanna-Maria N., Arja H., Mari K., Anna-Maija P. (2017). Collaborative partnership and the social value of clinical research: A qualitative secondary analysis. BMC Medical Ethics, 18(1), 1–12. 10.1186/s12910-017-0217-6 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Sloan R. A., Kim Y., Sahasranaman A., Muller-Riemenschneider F., Biddle S. J. H., Finkelstein E. A. (2018). The influence of a consumer-wearable activity tracker on sedentary time and prolonged sedentary bouts: Secondary analysis of a randomized controlled trial.(Report). BMC Research Notes, 11(1). 10.1186/s13104-018-3306-9 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Stickley T., Wright N., Slade M. (2018). The art of recovery: Outcomes from participatory arts activities for people using mental health services. Journal of Mental Health, 27(4), 367–373. 10.1080/09638237.2018.1437609 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Swanson J. M. (1986). Analyzing data for categories and description. In Chenitz W. C., Swanson J. M. (Eds.), From practice to grounded theory: Qualitative research in nursing (pp. 121–132). Addison-Wesley. [ Google Scholar ]
  • Szabo V., Strang V. R. (1997). Secondary analysis of qualitative data. Advances in Nursing Science, 20(2), 66–74. [ DOI ] [ PubMed ] [ Google Scholar ]
  • Ventola C. L. (2014). Mobile devices and apps for health care professionals: uses and benefits. Pharmacy and Therapeutics, 39(5), 356 https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4029126/pdf/ptj3905356.pdf [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Vorderstrasse A. A., Hammer M. J., Dungan J. R. (2014). Nursing implications of personalized and precision medicine. Seminars in Oncology Nursing, 30(2), 130 10.1016/j.soncn.2014.03.007 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Watters E. C., Cumming S., Caragata L. (2018). The lone mother resilience project: A qualitative secondary analysis. Forum: Qualitative Social Research, 19(2). 10.17169/fqs-19.2.2863 [ DOI ] [ Google Scholar ]
  • Westra B. L., Savik K., Oancea C., Choromanski L., Holmes J. H., Bliss D. (2011). Predicting improvement in urinary and bowel incontinence for home health patients using electronic health record data. Journal of Wound, Ostomy, and Continence Nursing, 38(1), 77. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Westra B. L., Sylvia M., Weinfurter E. F., Pruinelli L., Park J. I., Dodd D., Keenan G. M., Senk P., Richesson R.L., Baukner V., Cruz C., Gao G., Whittenburg L., Delaney C. W. (2017). Big data science: A literature review of nursing research exemplars. Nursing Outlook, 65(5), 549–561. 10.1016/j.outlook.2016.11.021 [ DOI ] [ PubMed ] [ Google Scholar ]
  • Ziebland S., Hunt K. (2014). Using secondary analysis of qualitative data of patient experiences of health care to inform health services research and policy. Journal of Health Services Research & Policy, 19(3), 177–182. 10.1177/1355819614524187 [ DOI ] [ PubMed ] [ Google Scholar ]
  • View on publisher site
  • PDF (473.1 KB)
  • Collections

Similar articles

Cited by other articles, links to ncbi databases.

  • Download .nbib .nbib
  • Format: AMA APA MLA NLM

Add to Collections

IMAGES

  1. (PDF) Normalizing Google Scholar data for use in research evaluation

    data analysis in research google scholar

  2. Revolutionizing Research: Google Scholar in 2024

    data analysis in research google scholar

  3. Data analysis in research

    data analysis in research google scholar

  4. Graph Showing Google Scholar Results by Year for the Following Search

    data analysis in research google scholar

  5. Significance Of Data Analysis In Academic Research

    data analysis in research google scholar

  6. Standard statistical tools in research and data analysis

    data analysis in research google scholar

COMMENTS

  1. Google Scholar

    Google Scholar provides a simple way to broadly search for scholarly literature. Search across a wide variety of disciplines and sources: articles, theses, books, abstracts and court opinions.

  2. An Overview of Data Analysis and Interpretations in Research

    Research is a scientific field which helps to generate new knowledge and solve the existing problem. So, data analysis is the crucial part of research which makes the result of the study more ...

  3. Learning to Do Qualitative Data Analysis: A Starting Point

    Yonjoo Cho is an associate professor of Instructional Systems Technology focusing on human resource development (HRD) at Indiana University. Her research interests include action learning in organizations, international HRD, and women in leadership. She serves as an associate editor of Human Resource Development Review and served as a board member of the Academy of Human Resource Development ...

  4. Analysing qualitative data

    Lee R, Fielding N. User's experiences of qualitative data analysis software. In: Kelle U, editor. Computer aided qualitative data analysis: theory, methods and practice. London: Sage; 1995. [Google Scholar] 11. Bloor M. On the analysis of observational data: a discussion of the worth and uses of inductive techniques and respondent validation.

  5. The Role of Data Analysis in Academic Research: Best Practices

    Data analysis ensures that the research findings are based on solid, objective evidence, contributing to the credibility of the study. Tip: Approach data analysis with an open mind, ready to discover unexpected trends and patterns that may alter your conclusions. 2. Choosing the Right Data Analysis Techniques. Different types of data require ...

  6. An Overview of the Fundamentals of Data Management, Analysis, and

    Quantitative research assumes that the constructs under study can be measured. As such, quantitative research aims to process numerical data (or numbers) to identify trends and relationships and to verify the measurements made to answer questions like who, how much, what, where, when, how many, and how. 1, 2 In this context, the processing of numerical data is a series of steps taken to help ...

  7. Approaches to Analysis of Qualitative Research Data: A Reflection on

    Few studies report on comparisons of the manual and technological approaches to qualitative data analysis. In one such study, Basit (2003) compares the use of the manual and technological approach to qualitative data analysis drawing on two research projects. She argues that the approach chosen is dependent on the size of the project, the funds ...

  8. Secondary Data Analysis: Using existing data to answer new questions

    Secondary data analysis is a valuable research approach that can be used to advance knowledge across many disciplines through the use of quantitative, qualitative, or mixed methods data to answer new research questions (Polit & Beck, 2021).This research method dates to the 1960s and involves the utilization of existing or primary data, originally collected for a variety, diverse, or assorted ...

  9. A General Inductive Approach for Analyzing Qualitative Evaluation Data

    A general inductive approach for analysis of qualitative evaluation data is described. The purposes for using an inductive approach are to (a) condense raw textual data into a brief, summary format; (b) establish clear links between the evaluation or research objectives and the summary findings derived from the raw data; and (c) develop a framework of the underlying structure of experiences or ...

  10. Secondary Data Analysis in Nursing Research: A Contemporary Discussion

    Given the widespread and growing use of secondary data analysis in scientific research, it is worthwhile revisiting this practice to appreciate how it can generate evidence in nursing, now and into the future. ... Qualitative Health Research, 7(3), 408-424. [Google Scholar] Hurlock-Chorostecki C., Forchuk C., Orchard C., Reeves S., van Soeren ...