Skip to main content

Core conflictual relationship theme: the reliability of a simplified scoring procedure

Abstract

Background

Creating a case formulation is an important and basic skill in psychotherapy meant to guide treatment. A patient’s interpersonal pattern is an essential part of a case formulation. Core Conflictual Relationship Theme (CCRT) is a well-known structured method to describe interpersonal patterns.

The CCRT method is based on the assumption that humans display a central relationship theme, which is shown in most relationships as well as in the patient-therapist relation. The CCRT scoring is based on how the patient describes interactions with others, in therapy sessions or in a specific interview. These descriptions are transcribed. Raters then score the identified relational episodes by choosing elements from the clustered categories of Wishes, Response from Others and Response from Self. The method has shown high validity and reliability. Inter rater reliability is generally good: Cohen’s kappa ranging from 0.55 to 0.70.

To decide CCRT pattern from transcribed material is time consuming and labour intensive This study investigates a labour- and timesaving version of the method.

Methods

This study aimed to investigate rater agreement in a simplified method of scoring the CCRT, based directly on live semi-structured dynamic interviews without transcribing the material.

Fifty-two patients referred for psychotherapy in a clinical trial, were scored for CCRT pattern. Based on information that came forth during the two-hour interview, raters scored the patients choosing elements from the clustered categories of Wishes, Response from Others and Response from Self. More than one category in each component could be chosen without ranking. Five raters compared two by two were investigated. Inter rater reliability was measured by Cohen’s kappa.

Results

Mean kappa for Wishes, Response from Others and Response from Self was .33, .41 and .45 respectively. Mean kappa for CCRT in total was .41 among 5 raters.

Conclusion

In this simplified method to score the CCRT based on oral dynamic interviews, fair to moderate IRR was obtained.

Trial registration

First Experimental Study of Transference-interpretations (FEST307/95).

Registration number: ClinicalTrials.gov Identifier: NCT00423462.

Peer Review reports

Background

Creating a case formulation is an important and basic clinical skill. Based on the patient’s narrative, it is a tool for the therapist to structure ideas about what has caused the patient’s problems, why the problems persist and how to make the patient feel better [1]. The case formulation can fill the gap between diagnosis and treatment and can be perceived as lying at the intersection of aetiology and description, theory and practice, and science and art [2]. When developed, the case formulation is based on the psychotherapist’s knowledge and preconceptions, what the patient has chosen to reveal, how a patient is presenting himself and interacts with the therapist. In this study a simplified method to describe patients is tested.

The primary function of a case formulation is to guide treatment. A psychodynamic formulation includes motivational components, aims to identify central dynamic patterns that repeat themselves over and across situations resulting in distress or psychological limitations, and guides clinicians in minimizing inferences from observable clinical data [3]. Many therapists have the opinion that case formulation is a time- consuming and difficult process, therefore, a simple mental construction would be sufficient.

Bowden [4] reported that 90% of educators in a study ranked case formulation skills as very important, even essential. However, case formulation skills are difficult to acquire. Kuyken and colleagues [5] studied the quality of case formulations constructed by 115 health professionals, 44% of the formulations were considered good enough. Eells and colleagues [6] evaluated 56 intake formulations, of which less than half described predisposing life events and/or inferred psychological mechanisms, as necessary in a proper case formulation.

Seitz [7] summoned a group of psychoanalysts in an effort to create reliable case formulations but they never succeeded. He concluded that they applied different levels of inference to the clinical data and hence never agreed on what was centrally important. A basic prerequisite in formulating cases is a certain level of agreement among raters. In their review, Barber and Crits-Christoph [8] found that structured case formulations can attain reliability. This is confirmed in a study by Garb [9] with clinicians sharing the same theoretical background. A recent publication also shows that very experienced clinicians with similar theoretical stance produced reliable, and thus clinically relevant formulations without elaborate instructions about how to structure the case formulations [10].

Psychotherapy is interpersonal in nature. Several methods have been developed over the last decades to identify and describe interpersonal patterns in psychotherapy, often as part of a case formulation. The most well-known is the Core Conflictual Relationship Theme (CCRT) method [11].

The CCRT method constructed by Luborsky and Crits-Christoph is based on the assumption that humans display a central relationship theme or transference pattern [12]. That is, people in general display the same patterns with different people, in various relationships, including the patient-therapist relationship [13]. In several studies this method has shown high reliability in identifying peoples wishes, responses from others and responses of self [11, 14]. There is also evidence for the validity of the CCRT [15, 16], [14]. The CCRT has strong convergent validity with other transference- related measures such as the circumplex model of human reactions in relation to others, the Structural Analysis of Social Behavior (SASB) [17, 18].

The CCRT scoring is based on narratives presented by the patient [19]. The narratives should describe interactions with others, even the therapist. These descriptions are defined as relational episodes (REs), which includes three components: the patient’s expectations or wishes (W) in meeting another person, how the other is considered or expected to react (RO; Response Other) and how the patient responded (RS; Response Self). REs depict real, imagined or dreamt episodes from the patient’s point of view. An RE told by the patient, can be more or less complete. The REs can be collected from therapy sessions or directly from an interview focusing on relationships with central others, as in the Relational Anecdote Paradigm (RAP) [20] (pp 109–120) developed by Crits-Christoph and Luborsky.

In the CCRT method, trained scorers identify and demarcate REs in the transcribed material from therapy sessions or the RAP interview. Other trained independent scorers, not involved in the identification of REs in the transcriptions, identify and categorise W, RO, and RS from the demarcated REs. The most frequent pattern is considered the most useful description of a patient’s CCRT. The goal is to avoid interpreting the patient’s narrative. The chosen categories should reflect literally what the patient told, and inferences should be avoided.

An example of an RE and a tailor-made description of W, RO and RS can illustrate the process. Here is an example of an identified RE:

“He came over, unannounced to have a coffee with me. I pretended to enjoy his visit, since he is a friend since long … but actually I really wanted to sit down and read my book, or rather this was keeping me from reading and that hassled me. I really resented it a lot. With a guy like this, he has helped me a lot before for which I’m grateful. He’s just in his own world … insensitive to others’ needs. And you know he wouldn’t understand if I told him. He would be so sad; you know it was kind of a hassle”. The episode is complete. The tailor-made descriptions can be described as follows: W- to be free of unwanted visitor. RO - “He wouldn’t understand, he would be so sad” and RS - I feel hassled, resentful, guilty and compelled to suffer his presence.

The CCRT standard categories have been empirically chosen from the most frequently used ones, resulting in a standard category list. The third edition is the most widely used. Cluster analysis of the lists of categories resulted in the creation of “clustered standard categories” consisting of eight each for the different components, W, RO and RS [20] (pp. 43–54). From the cited quotation above the following clustered standard categories can be chosen: W: To be distant and avoid conflicts, RO: Upset, RS: Oppose and hurt others (see Table 2).

In an effort to fit in a CCRT pattern into a circumplex model of interpersonal patterns, Crits Christoph et al. [21] developed Quantitative Assessment of Interpersonal Themes (QUAINT). They tried to organise the clusters according to Benjamin’s [22] SASB. A study [23] ended up in 30 W, 31 RO and 40 RS. Most categories of interpersonal behaviour could be rated reliably. However, the study showed poor interrater reliability in some items, mostly among negative wishes.

The original CCRT scoring method is labour intensive and time consuming. First, researchers transcribe therapy sessions or RAP interviews, then independent judges identify REs, and other judges identify the categories in each component and count them to determine the most frequent categories. The categories are ranked according to frequency.

Luborsky and Crits-Christoph’s [11] original work showed an agreement between two raters, scoring 35 cases and reported weighted kappa values ranging from 0.61 to 0.70 in pair comparisons. Since then many studies have been performed with varying levels of reliability. See some examples in Table 1.

Table 1 Kappa in CCRT Reliability Studies

Zander et al. [25] scored CCRT directly from a videotaped interview. The results were presented as unweighted kappa, with the following components: W = 0.35, RO = 0.41, RS = 0.46; The study has been criticised by Luborsky and Diguer [24], partly because of its use of kappa instead of weighted kappa which is used in most of the reliability studies on CCRT.

The results indicate that the original CCRT method can be used in research, but it is probably too labour intensive to be used in ordinary clinical practice, supervision or education.

Aim

The traditional CCRT method has shown itself useful in developing case formulations. Though, it is elaborate and time consuming. The present study aimed to test a simplified method. We investigated rater agreement in a method to establish a CCRT pattern directly from a Dynamic Interview (DI) without transcribing the material.

Methods

This study is part of the First Experimental Study of Transference-interpretations (FEST). The FEST is a dismantling randomised clinical trial designed to study the long-term effects of transference interpretations in psychodynamic psychotherapy [28].

Patients

The patients were referred to the study therapists by primary care physicians, private specialist practices and public outpatient departments. No patient was directly recruited for research.

The patients were offered exploratory psychotherapy due to depressive disorders, anxiety disorders, personality disorders and interpersonal problems. Patients with psychosis, bipolar disorder, organic mental disorder or substance abuse were excluded. Written informed consent was obtained from each participant.

After taking history and assessment of background variables by the patient’s therapists, an independent evaluator interviewed each patient in a two-hour psychodynamic semi-structured interview DI, modified after Malan and Sifneos [28]. The patients were asked to talk about themselves and the interviewer should seek to elucidate the dynamics behind the patient’s beliefs, affective experiences, behaviour, maladaptive/adaptive relationships and symptoms. The patients were helped to explore meaningful experiences and vignettes in detail and to give examples of interaction with others in their life. If necessary, they were asked more detailed questions in a manner that illuminated their personal characteristics. When clarifications were used, they were made without elaborations or inferences. In the interview there were no specific focus on the CCRT components.

If patients in about the middle of the interview hadn’t come up with enough information the interviewer asked specifically about parents and friends. The patients were asked to describe their relationships and give concrete examples. Each interview was audiotaped.

Therapists and evaluators

The patients were assigned to one of seven therapists, depending on the latter’s availability. The therapists comprised six psychiatrists and one clinical psychologist. They had experiences in practising psychodynamic psychotherapy for 10 to 25 years. Four of them were psychoanalysts. The therapists each treated 10 to 17 patients in the study.

Evaluation and CCRT-scoring

In the FEST the DI [28], was primarily used to obtain the scores on the Psychodynamic Functioning Scales (PFS). The version of the CCRT in this study used the eight standard cluster categories in each of the three components W, RO and RS. The raters scored 58 of the patient interviews The W standard categories used in this study were: To assert self and be independent, To oppose, hurt and control others, To be controlled, hurt, and not responsible, To be distant and avoid conflicts, To be close and accepting, To be loved and understood, To feel good and comfortable, To achieve and help others. The RO categories used were: Strong and independent, Controlling, Upset, Bad, Rejecting and opposing, Helpful, Likes me, Understanding. The RS categories used were: Helpful, Unreceptive, Respected and accepted, Oppose and hurt others, Self-controlled and self-confident, Helpless, Disappointed and depressed, Anxious and shameful.

In all the 58 interviews the information that came forth, was considered sufficient to make CCRT- scoring possible. The scorings were made after listening to the whole interview based on a global impression. One interview was scored by all raters. Most of the interviews were scored by three or four raters (39 patients). The raters were numbered from 1 to 7. The most active raters were number 1 (54 patients), 2 (46 patients), 6 (40 patients), and 7 (34 patients).

In this study more than one category (e.g. controlling and upset) in each component (e.g. response of other) could be chosen, without ranking. This differed from the original method where the frequencies of the categories were registered and ranked according to frequency.. In the FEST, only pairs of raters scoring the same 15 or more patient interviews at a specific time point were included in the analyses. Then the scores presented by the raters were compared in each component (W, RO and RS) respectively.

Statistics

Cohen’s kappa statistics [29] were used. This method has become the standard for scoring agreement on the CCRT [20] as it takes into consideration agreement occurring by chance.

The data were analysed using SPSS [30] IRR was investigated by comparing every possible combination of pairs of raters, scoring the same patient. Landis and Koch [31], have arbitrarily defined intervals in Cohen’s kappa for inter rater agreement, where 0–0.20 is considered as slight, 0.21–0.40 as fair, 0.41–0.60 as moderate, 0.61–0.80 as substantial, and 0.81–1 as almost perfect agreement.

In the original method raters were instructed to note the frequencies of different categories in each component. When both raters listed the same W, RO or RS as the most frequent, it was given the highest weight. A lower weight was given when one judge’s most frequent score matched the second highest score from the other judge. The lowest weight was given when the two judges’ second highest scores matched. That method made it possible to use weighted kappa.

In this study, kappa was used as there were only categorical data without mutual ranking. The evaluators were never asked to rank the suitability of the chosen categories. Where two raters both scored the same category, it was indicated as 1 and when one or the other scored a category, nor scored by the other it was indicated as 0. Therefore, no weighted kappa could be presented. The kappa was compared between individual raters scoring as in the original study.

Results

The raters scored one or more categories in each main component (W/RO/RS) without ranking the categories. As shown in Table 1, also in this study, Cohen’s kappa differed among different raters. In total, the results varied from kappa 0.60 to 0.26. The mean values were as follows: W = 0.33 (0.26–0.52), RO = 0.44 (0.29–0.60) and RS = 0.45 (0.31–0.58). CCRT in total: Mean kappa: 0.41. All results were significant.

Discussion

This study aimed to investigate the reliability in a simplified method of scoring CCRT by calculating IRR among raters’ scoring based on a DI. This less time-consuming method could make CCRT more applicable in psychotherapy research. However, scoring based on the information gathered from an interview used in ordinary clinical practice, makes it difficult to obtain IRR as high as in the original method. Although experienced therapists with considerable training scored the interviews, the results showed an overall fair to moderate IRR (Table 2). DIs were used in this investigation, primarily to score other instruments, but the same interviews were used scoring the CCRT. It was given lesser attention among the raters as it was considered a possible spin off effect of the interview.

Table 2 CCRT, inter rater reliability

The ratings on RO and RS showed the highest level of agreement and the W ratings indicated the lowest. These findings, with the least degree of agreement in W, are comparable with those reported by Crits- Christoph and Luborsky [20]. Possibly, this is a sign of raters’ interpreting more, having difficulty to score wishes literally.

In the CCRT method the most frequent category was the one to be chosen, but the second and the third most frequent were also identified, which made weighted data possible [20]. In this study, the raters were instructed to score the observed dominating category (categories), presented in each interview, without any rank. Hence, weighted kappa could not be used.

The raters were asked to score the CCRT from the global impression obtained in the Dynamic interview. Typically, the raters scored four to five categories in total for each patient. The choice to score more than the most dominating, or the most frequent category, was debated by Luborsky and Crits Christoph, who found it more informative to only use the most frequent category in each component [20]. Other studies using CCRT used the opportunity to score more than one category in the different components to declare a more complete description of a patient’s relational patterns [32].

In the studies presented in Table 1, the assessments of IRR vary. Some studies use weighted kappa and some kappa depending on the scoring method. In the present study, there were different numbers of raters scoring the same patient. Additionally, they had the possibility to choose more than one category, without ranking, in each component. This necessarily increased the variance in the results, causing lower kappa values. As noted above, kappa was shown in the present study, whereas weighted kappa is more permissive of varying measures, and therefore as a rule, presents higher kappa values. A previous study showed a weighted kappa for the CCRT pattern in total [26] while the others (see Table 1), showed kappa or weighted kappa for the separate components. In the present study the mean kappa value for the total CCRT pattern resulted in moderate reliability (0.41). Since categorical data were used in this study, its results are not directly comparable with those of Luborsky and Crits Christoph.

A limitation involved the varying numbers of raters scoring the patients. More consistent results might have been obtained with a constant number of raters. Nonetheless, as the raters were compared with one another, and only pairs rating 15 or more patients in common were included in the analyses this limitation presented is a lesser problem.

Scoring more than one item e.g. in W, could increase the chance that one rater is unanimous with another rater in a pair of compared raters, but as Cohen’s kappa is used, if the two raters disagree on another category in the same component, either if one rater just scored one category (and hence disagree) or a different category, it will result in a lower kappa. With ranked categories and the use of weighted kappa, results more similar to other studies might have been obtained.

A strength was that relatively many patients were included in this study compared with other studies investigating IRR in the CCRT. The number of raters in this study could also be considered adequate compared with those in other studies investigating reliability in the CCRT scoring. The raters were experienced clinicians and had participated in earlier research and were skilled in scoring psychometric measures. This might have resulted in higher internal validity but perhaps at the expense of lesser external validity.

The use of kappa as a method to investigate IRR in medical sciences has been criticised, as being far too accepting of low rater agreement considered to be good enough IRR [33]. For example, this issue can be understood when looking at for instance, evaluating a diagnosis of cancer by microscopy, where the raters have to choose “yes” or “no”, as using few categories as a rule results in higher kappa values [34].

In a study performed by Zander and colleagues [25], varying reliability results (0.35–0.46), presented as kappa values, were shown when assessing CCRT directly from videotaped RAP interviews. The present study suggests that at least fair to moderate reliability can be achieved when using a more time-saving procedure, even with interviews primarily used to score other psychometric instruments. The component W as a rule is more difficult to attain reliability in and in this study only reach fair IRR, it might be doubtful to use in research. The other components, RO and RS can possibly be used as they in this study met moderate IRR.

An unanswered question is whether a simplified CCRT scoring based directly on RAP interview, with a focus on REs, without transcribing the interviews, could have improved IRR although this was not the case Zander and colleagues’ study [25] using the German version of the CCRT. In this study a Dynamic interview was used to establish a CCRT pattern. Probably the same interview with more focus on the CCRT and questions to probe for the different components in the CCRT could yield improved agreement. Maybe this simplified scoring method can make it useful in ordinary clinical settings and education.

Conclusions

This study aimed to investigate IRR in a simplified method of scoring the CCRT based directly on a DI without transcribing the material. The IRR analyses showed fair to moderate rater agreement in a highly controlled research context. Based on this study, RO and RS could be used for research purposes. This study implicates that it is possible to score the CCRT directly from an interview with acceptable IRR.

Availability of data and materials

The datasets generated and/or analysed during the current study are not publicly available due to privacy but are available from the corresponding author on reasonable request.

Abbreviations

CCRT:

Core Conflictual Relationship Theme

DI:

Dynamic interview

FEST:

First Experimental Study of Transference–interpretations

IRR:

Inter rater reliability

RAP:

Relationship anecdotal paradigm

RE:

Relational episode

RO:

Response from others

RS:

Response from self

W:

Wishes to others

References

  1. 1.

    Eells T. Handbook of psychotherapy case formulation. New York, London: The Guilford Press; 1997.

    Google Scholar 

  2. 2.

    Sim K, Gwee KP, Bateman A. Case formulation in psychotherapy: revitalizing its usefulness as a clinical tool. Acad Psychiatry. 2005;29:289–92.

    Article  Google Scholar 

  3. 3.

    Hoglend P, Guldberg CA, Perry JC. Scientific approaches to making psychodynamic formulations. Nord J Psychiatry. 1992;46:41–8.

    Article  Google Scholar 

  4. 4.

    Bowden CL, Humphrey FJ, Thompson MG. Priorities in psychiatric residency training. Am J Psychiatry. 1980;137:1243–6.

    CAS  Article  Google Scholar 

  5. 5.

    Kuyken W, Fothergill CD, Musa M, Chadwick P. The reliability and quality of cognitive case formulation. Behav Res Ther. 2005;43:1187–201.

    Article  Google Scholar 

  6. 6.

    Eells TD, Kendjelic EM, Lucas CP. What’s in a case formulation? Development and use of a content coding manual. J Psychother Pract Res. 1998;7:144–53.

    CAS  PubMed  PubMed Central  Google Scholar 

  7. 7.

    Seitz P. The consensus problem in psychoanalytical research. Des Moines: Meredith Publishing Company; 1966.

  8. 8.

    Barber JP, Crits-Christoph P. Advances in measures of psychodynamic formulations. J Consult Clin Psychol. 1993;61:574–85.

    CAS  Article  Google Scholar 

  9. 9.

    Garb HN. Studying the clinician: judgment research and psychological assessment. Washington, DC: American Psychological Association; 1998. p. 333.

    Google Scholar 

  10. 10.

    Sørbye Ø, Dahl HS, Amlo S, Hersoug AG, Haukvik U, Hartberg CB, et al. Scientific approaches to making psychodynamic formulations. Nord J Psychiatry. 1992;46(1):41 In press.

    Article  Google Scholar 

  11. 11.

    Crits-Christoph P, Luborsky L, Dahl L, Popp C, Mellon J, Mark D. Clinicians can agree in assessing relationship patterns in psychotherapy. The Core conflictual relationship theme method. Arch Gen Psychiatry. 1988;45:1001–4.

    CAS  Article  Google Scholar 

  12. 12.

    Freud S. The dynamics of transference. Freud Complete Works (1890-1939). 1912. p. 1034–38. https://www.valas.fr/IMG/pdf/Freud__complete works.pdf.

  13. 13.

    Sandler J, Holder A, Kawenoka M, Kennedy H, Neurath L. Notes on some theoretical and clinical aspects of transference. Int J Psychoanal. 1969;50:633–45.

    Google Scholar 

  14. 14.

    Barber JP, Luborsky L, Diguer L, Crits-Christoph P. A comparison of core conflictual relationship themes before psychotherapy and during early sessions. J Consult Clin Psychol. 1995;63:145–8.

    CAS  Article  Google Scholar 

  15. 15.

    Levine FJ, Luborsky L. Object and self. A developmental approach. New York: International Universities Press; 1981.

    Google Scholar 

  16. 16.

    Crits-Christoph P, Cooper A, Luborsky L. The accuracy of therapists’ interpretations and the outcome of dynamic psychotherapy. J Consult Clin Psychol. 1988;56:490–5.

    CAS  Article  Google Scholar 

  17. 17.

    Henry WP, Schacht TE, Strupp HH. Structural analysis of social behavior: application to a study of interpersonal process in differential psychotherapeutic outcome. J Consult Clin Psychol. 1986;54:27–31.

    CAS  Article  Google Scholar 

  18. 18.

    Schacht TE, Binder JL, Strupp HH. The dynamic focus. In: Strupp HH, Binder JL, editors. Psychotherapy in a new key: a guide to time-limited dynamic psychotherapy. New York: Basic Books; 1984.

    Google Scholar 

  19. 19.

    Luborsky L, Crits-Christoph P, Mellon J. Advent of objective measures of the transference concept. J Consult Clin Psychol. 1986;54:39–47.

    CAS  Article  Google Scholar 

  20. 20.

    Luborsky L, Crits-Christoph P. Understanding transference the Core conflictual relationship theme, vol. 379. Washington: American Psychologica Association; 1998.

    Google Scholar 

  21. 21.

    Crits-Christoph P, Demorest A, Muenz LR, Baranackie K. Consistency of interpersonal themes for patients in psychotherapy. J Pers. 1994;62:499–526.

    CAS  Article  Google Scholar 

  22. 22.

    Benjamin LS. Structural analysis of social behaviour. Pschological Review. 1974;81:392–425.

  23. 23.

    Crits-Christoph P, Connolly MB, Shaffer C. Reliability and base rates of interpersonal themes in narratives from psychotherapy sessions. J Clin Psychol. 1999;55:1227–42.

    CAS  Article  Google Scholar 

  24. 24.

    Luborsky L, Diguer L. A novel reliability study: reply to Zander et al. Psychother Res. 1995;5(3):237–41.

    Article  Google Scholar 

  25. 25.

    Zander B, Strack M, Cierpka M, Reich G, Staats H. Coder agreement using the German edition of Luborsky’s CCRT method in videotaped or transcribed RAP interviews. Psychother Res. 1995;5(3):231–6.

    Article  Google Scholar 

  26. 26.

    Luborsky L, Diguer L, Andrusyna T, et al. A method of choosing CCRT scorers. Psychother Res. 2004;14:127–34.

    Article  Google Scholar 

  27. 27.

    Hegarty BD, Marceau EM, Gusset M, Grenyer BFS. Early treatment response in psychotherapy for depression and personality disorder: links with core conflictual relationship themes. Psychother Res. 2019;1:1–12.

    Google Scholar 

  28. 28.

    Hoglend P, Amlo S, Marble A, et al. Analysis of the patient-therapist relationship in dynamic psychotherapy: an experimental study of transference interpretations. Am J Psychiatry. 2006;163:1739–46.

    Article  Google Scholar 

  29. 29.

    Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas. 1960;20:1.

    Article  Google Scholar 

  30. 30.

    IBM Corp. Released 2015. IBM SPSS statistics for windows VA. New York: IBM CorpSPSS Statistics version 23.0; 2015.

    Google Scholar 

  31. 31.

    Landis JR, Koch GG. An application of hierarchical kappa-type statistics in the assessment of majority agreement among multiple observers. Biometrics. 1977;33:363–74.

    CAS  Article  Google Scholar 

  32. 32.

    Wiseman H, Tishby O. Applying relationship anecdotes paradigm interviews to study client-therapist relationship narratives: Core conflictual relationship theme analyses. Psychother Res. 2017;27:283–99.

    Article  Google Scholar 

  33. 33.

    McHugh ML. Interrater reliability: the kappa statistic. Biochem Med (Zagreb). 2012;22:276–82.

    Article  Google Scholar 

  34. 34.

    Maclure M, Willett WC. Misinterpretation and misuse of the kappa statistic. Am J Epidemiol. 1987;126:161–9.

    CAS  Article  Google Scholar 

Download references

Acknowledgements

The authors thank, Svein Amlo, MD; Kjell Petter Bøgwald, MD, PhD; Øystein Sørbye, MD; Oscar Heyerdahl, MD; Alice Marble, PsyD; and Mary Cosgrove Sjaastad, MD for their contribution in peer supervision, the development of research questions and the decisions of outcome measures, as well as for providing the treatment data for the study. They are all psychotherapists in private practice. The authors express their gratitude to all the participating patients for their central role in this investigation and good cooperation in follow-up studies.

Patient involvement

All the participating evaluators and therapists have themselves experience as patients in psychotherapy. They have collaborated in developing the research questions and outcome measures, as well as in designing the study. All the results have in part been disseminated to the patients in the study. Concerning this small part of the study, all data about the specific patients are deleted, no personal information about the results can be delivered.

The patients in the FEST were ordinary patients in the mental health care system in Helse Sør- Øst. Each received an extensive expert evaluation and assessment and a more extensive treatment than regular therapy. It may have been inconvenient for some patients to attend in-group interviews and fill out a lot of self-reports. Although some patients described the follow-up interviews as constituting a kind of “booster” therapy. No dangers or drawbacks were indicated for the participants in this study. All patients appeared in person for the three-year follow-up interviews.

Funding

This study was funded by the Division of Mental Health and Addiction, University of Oslo and Østfold Hospital Trust.

Author information

Affiliations

Authors

Contributions

RU is the second author and main supervisor of this study. She is a researcher in the FEST-research group with the responsibility for micro-process analyses. PH is the principal investigator in the FEST. PH and H-SJD, participated in providing and analysing the treatment data, as well as drafting and critically revising the manuscript for important intellectual content. All authors have given their final approval of the manuscript.

Corresponding author

Correspondence to Peter Tallberg.

Ethics declarations

Ethics approval and consent to participate

The Regional Ethics Committee for Health Region 1 in Norway approved the study protocol and the information given to the patients in the First Experimental Study of Transference- interpretations (FEST307/95). Written informed consent was obtained from each participant.

Consent for publication

The ethical approval and the consent to participate included the consent to publish. The patient material and the data collected including case material were accepted for use in research and publishing as well as in teaching.

Competing interests

The authors declare no financial or non-financial competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Tallberg, P., Ulberg, R., Johnsen Dahl, HS. et al. Core conflictual relationship theme: the reliability of a simplified scoring procedure. BMC Psychiatry 20, 150 (2020). https://doi.org/10.1186/s12888-020-02558-4

Download citation

Keywords

  • Relationship patterns
  • Psychodynamic
  • Core conflictual relationship theme
  • Rater agreement