MIT Libraries home DSpace@MIT

  • DSpace@MIT Home
  • MIT Libraries
  • Doctoral Theses

Learning task-specific similarity

Thumbnail

Other Contributors

Terms of use, description, date issued, collections.

IT-University of Copenhagen Logo

  • Support to ITU faculty

Algorithms for Similarity Search and Pseudorandomness

  • Computer Science

Research output : Book / Anthology / Report / Ph.D. thesis › Ph.D. thesis

Access to Document

  • PhD Thesis Final Version Tobias Lybecker Christiani Final published version, 1.38 MB

Fingerprint

  • Nearest neighbor search Engineering & Materials Science 100%
  • Hash functions Engineering & Materials Science 35%
  • Random access storage Engineering & Materials Science 16%
  • Tuning Engineering & Materials Science 11%
  • Set theory Engineering & Materials Science 9%

T1 - Algorithms for Similarity Search and Pseudorandomness

AU - Christiani, Tobias Lybecker

N2 - We study the problem of approximate near neighbor (ANN) searchand show the following results:• An improved framework for solving the ANN problem usinglocality-sensitive hashing, reducing the number of evaluationsof locality-sensitive hash functions and the word-RAM complexity compared to the standard framework.• A framework for solving the ANN problem with space-timetradeoffs as well as tight upper and lower bounds for the spacetime tradeoff of framework solutions to the ANN problemunder cosine similarity.• A novel approach to solving the ANN problem on sets alongwith a matching lower bound, improving the state of theart. A self-tuning version of the algorithm is shown throughexperiments to outperform existing similarity join algorithms.• Tight lower bounds for asymmetric locality-sensitive hashingwhich has applications to the approximate furthest neighborproblem, orthogonal vector search, and annulus queries.• A proof of the optimality of a well-known Boolean localitysensitive hashing scheme.We study the problem of efficient algorithms for producing highquality pseudorandom numbers and obtain the following results:• A deterministic algorithm for generating pseudorandom numbers of arbitrarily high quality in constant time using nearoptimal space.• A randomized construction of a family of hash functions thatoutputs pseudorandom numbers of arbitrarily high qualitywith space usage and running time nearly matching knowncell-probe lower bounds.

AB - We study the problem of approximate near neighbor (ANN) searchand show the following results:• An improved framework for solving the ANN problem usinglocality-sensitive hashing, reducing the number of evaluationsof locality-sensitive hash functions and the word-RAM complexity compared to the standard framework.• A framework for solving the ANN problem with space-timetradeoffs as well as tight upper and lower bounds for the spacetime tradeoff of framework solutions to the ANN problemunder cosine similarity.• A novel approach to solving the ANN problem on sets alongwith a matching lower bound, improving the state of theart. A self-tuning version of the algorithm is shown throughexperiments to outperform existing similarity join algorithms.• Tight lower bounds for asymmetric locality-sensitive hashingwhich has applications to the approximate furthest neighborproblem, orthogonal vector search, and annulus queries.• A proof of the optimality of a well-known Boolean localitysensitive hashing scheme.We study the problem of efficient algorithms for producing highquality pseudorandom numbers and obtain the following results:• A deterministic algorithm for generating pseudorandom numbers of arbitrarily high quality in constant time using nearoptimal space.• A randomized construction of a family of hash functions thatoutputs pseudorandom numbers of arbitrarily high qualitywith space usage and running time nearly matching knowncell-probe lower bounds.

M3 - Ph.D. thesis

SN - 978-87-7949012-3

T3 - ITU-DS

BT - Algorithms for Similarity Search and Pseudorandomness

PB - IT-Universitetet i København

similarity search phd thesis

Plagiarism and what are acceptable similarity scores?

Dec 1, 2020 • knowledge article, information.

The Similarity Report is a flexible document that provides a summary of matching or similar text in submitted work compared against a huge database of Internet sources, journals and previously submitted work, allowing students and instructors to review matches between a submitted work and the database scanned by Turnitin. Therefore, the Turnitin Similarity Report does not define whether or not a student's work is plagiarized. The instructor responsible for the course - as a subject matter expert - has a duty to exercise academic judgement on the work that is submitted to Turnitin for their classes. The percentage that is returned on a student's submission (called similarity index or similarity score) defines how much of that material matches other material in the database, it is not a marker as to whether a student has or has not plagiarized. Matches will be displayed to material that has been correctly cited and used, which is where the instructor's academic judgement must come into play. Please find our guide links below on how to interpret the Similarity Report and its similarity score: If you are a student, click here . If you are an instructor, click here . 

  • Copyright © 2024 Turnitin, LLC. All rights reserved.
  • Turnitin.com
  • Release Notes
  • Known Issues
  • Privacy and Security
  • System Status

Learning Task-Specific Similarity

The right measure of similarity between examples is important in many areas of computer science. In particular it is a critical component in example- based learning methods. Similarity is commonly defined in terms of a conventional distance function, but such a definition does not necessarily capture the inherent meaning of similarity, which tends to depend on the underlying task. We develop an algorithmic approach to learning similarity from examples of what objects are deemed similar according to the task-specific notion of similarity at hand, as well as optional negative examples. Our learning algorithm constructs, in a greedy fashion, an encoding of the data. This encoding can be seen as an embedding into a space, where a weighted Hamming distance is correlated with the unknown similarity. This allows us to predict when two previously unseen examples are similar and, importantly, to efficiently search a very large database for examples similar to a query.

This approach is tested on a set of standard machine learning benchmark problems. The model of similarity learned with our algorithm provides and improvement over standard example-based classification and regression. We also apply this framework to problems in computer vision: articulated pose estimation of humans from single images, articulated tracking in video, and matching image regions subject to generic visual similarity.

Thesis chapters

  • Front matter (of little scientific interest)
  • Chapter 1: Introduction

This chapter defines some technical concepts, most importantly the notion of similarity we want to model, and provides a brief overview of the contributions of the thesis.

  • Chapter 2: Background

Among the topics covered in this chapter: example-based classification and regression, previous work on learning distances and (dis)similarities (such as MDS), and algorithms for fast search and retrieval, with emphasis on locality sensitive hashing (LSH).

  • Chapter 3: Learning embeddings that reflect similarity
  • Similarity sensitive coding (SSC). This algorithm discretizes each dimension of the data into zero or more bits. Each dimension is considered independently of the rest.
  • Boosted SSC. A modification of SSC in which the code is constructed by greedily collecting discretization bits, thus removing the independence assumption.

The underlying idea of all three algorithms is the same: build an embedding that, based on training examples of similar pairs, maps two similar objects close to each other (with high probability). At the same time, there is an objective to control for "spread": the probability of arbitrary two objects (in particular of dissimilar pairs of objects, if examples of such pairs are available) to be close in the embedding space should be low.

This chapter also describes results of an evaluation of the proposed algorithms on seven benchmarks data sets from UCI and Delve repositories.

  • Chapter 4: Articulated pose estimation

An application of the ideas developed in previous chapters to the problem of pose estimation: inferring the articulated body pose (e.g. the 3D positions of key joints, or values of joint angles) from a single, monocular image containing a person.

  • Chapter 5: Articulated tracking

In a tracking scenario, a sequence of views, rather than a single view of a person, is available. The motion provides additional cues, which are typically used in a probabilistic framework. In this chapter we show how similarity-based algorithms have been used to improve accuracy and speed of two articulated tracking systems: a general motion tracker and a motion-driven animation system focusing on swing dancing.

  • Chapter 6: Learning image patch similarity

An important notion of similarity that is naturally conveyed by examples is the visual similarity of image regions. In this chapter we focus on a particular definition of such similarity, namely invariance under rotation and slight shift. We show how the machinery developed in Chapter 3 allows us to improve matching performance for two popular representations of image patches.

  • Chapter 7: Conclusions
  • Bibliography

Go to the homepage

  • Apps & tools
  • Library access browser extension
  • Readspeaker Textaid
  • Access & accounts
  • Accessibility tools

iThenticate – Similarity check for researchers

  • Keylinks Learning Resources
  • Working with courses
  • Faculty support
  • Canvas tools
  • Integrated third party tools
  • Manuals and videos
  • FAQ for teachers and tutors
  • Canvas FAQ for students
  • Open Access Journal Browser
  • Qualtrics survey tool
  • Remote access to licensed resources and software
  • Virtual Research Environment – VRE
  • Wooclap for interaction
  • Zoom Videoconferencing tool
  • Video editing tools
  • Video recording tools
  • Check for software

Why and what?

Maastricht University endorses the principles of scientific integrity and therefore provides services to check for the similarity between documents. Separate services are provided for research and educational purposes.

Every UM-affiliated researcher can use this service. Ithenticate – provided by TurnItIn – compares your submitted work to millions of articles and other published works and billions of webpages.

Check a manuscript / PhD thesis

This tool is for research purposes only!

Not for educational purposes

The Similarity Check Service is not intended for educational purposes (e.g., checking master’s theses for plagiarism). Please use Turnitin Originality instead (available through the digital learning environment Canvas). Turnitin Originality is tailored to the specific requirements for educational purposes.

The maximum number of submissions for these services is adapted to their respective purposes.

Support & Contact

In case you are in doubt about which similarity check service to use for a particular purpose, please contact us so we can find a suitable solution for you while guaranteeing the sustainable availability of the services for all UM scholars.

Plagiarism and how to prevent it

Plagiarism is using someone else’s work or findings without stating the source and thereby implying that the work is your own. When using previously established ideas that add pertinent information in a research paper, every researcher should be cautious not to fall into the trap of sloppy referencing or even plagiarism.

Plagiarism is not just confined to papers, articles or books, it can take many forms (for more information, see this infographic by iThenticate ).

The Similarity Check Service can help you to prevent only one type of plagiarism: verbatim plagiarism, and only if the source is part of the corpus.

The software does not automatically detect plagiarism; it provides insight into the amount of similarity in the text between the uploaded document and other sources in the corpus of the software. This does not mean this part of the text is viewed as plagiarism in your specific field. For instance, the methods section in some subfields follows very common wording, which could lead to a match. If there are instances where the submission’s content is similar to the content in the database, it will be flagged for review and should be evaluated by you.

How to use the service

Getting started.

Go to iThenticate and enter your UM username and password in the appropriate fields. Select ‘login’.

Library Access - login

2. First-time user

As a first-time user, you will then have to check your personal information and declare that you agree to the Terms and conditions.

similarity search phd thesis

3. My Folders and My Documents iThenticate

iThenticate will provide you with a folder group My Folders and a folder within that group titled My Documents.

similarity search phd thesis

From the My Documents folder, you will be able to submit a document by selecting the Submit a document link.

similarity search phd thesis

4. Upload a file

On the Upload a file page, enter the authorship details and the document title. Select Choose File and locate the file on your device.

similarity search phd thesis

Select the Add another file link to add another file. You can add up to ten files before submitting. Select Upload to upload the document(s).

5. Similarity Report

To view the Similarity Report for the paper, select the similarity score in the Report column. It usually takes a couple of minutes for a report to generate.

similarity search phd thesis

Finding your way around

The main navigation bar at the top of the screen has three tabs. Upon logging in, you will automatically land on the folders page.

similarity search phd thesis

This is the main area of iThenticate. From the folders page, you will be able to upload, manage and view documents.

The settings page contains configuration options for the iThenticate interface.

Account Info

The account information page contains the user profile and account usage.

Options for exclusion

There can be various reasons why you may want to exclude certain sources that your document is compared to or certain parts of your document in the similarity check. You can specify options for exclusion in the Folder settings.

similarity search phd thesis

If you choose to exclude ‘small matches’, you will be asked to specify the minimum number of words that you want to be shown as a match.

If you choose to exclude ‘small sources’, you will be asked to specify a minimum number of words or a minimum match percentage.

Once you click Update Settings, the settings will be applied to the particular folder.

Manuals & training videos

iThenticate provides a scale of up-to-date manuals and instructions on their own website. Please consult them here .

You can also use these training videos to learn how to use the service.

Please be aware that information in these manuals and videos about logging in and account settings are not applicable to UM users of this service.

How to read the similarity report

The similarity report provides the percentage of similarity between the submitted document and content in the iThenticate database. This is the type of report that you will use most often for a similarity check.

It is perfectly natural for a submitted document to match against sources in the database, for example if you have used quotes. 

The similarity score simply makes you aware of potential problem areas in the submitted document. These areas should then be reviewed to make sure there is no sloppy referencing or plagiarism.

iThenticate should be used as part of a larger process, in order to determine if a match between your submitted document and content in the database is or is not acceptable.

This video shows how to read the various reports

This video shows how the Document viewer works  

Academic Integrity and Plagiarism

Everyone involved in teaching and research at Maastricht University shares in the responsibility for maintaining academic integrity (see Scientific Integrity ). All academic staff at UM are expected to adhere to the general principles of professional academic practice at all times.

Adhering to those principles also includes preventing sloppy referencing or plagiarism in your publications.

Additional information on how to avoid plagiarism can be also be found in the Copyright portal of the library.

Sources used

iThenticate compares the submitted work to 60 million scholarly articles, books, and conferences proceedings from 115,000 scientific, technical, and medical journals, 114 million Published works from journals, periodicals, magazines, encyclopedias, and abstracts, 68 billion current and archived web pages.

Checking PhD theses

The similarity check service (iThenticate) can be used by doctoral candidates or their supervisors to assess the work. Find out about the level of similarity with other publications and incorrect referencing before you send (parts of) the thesis to the Assessment Committee, a publisher or send in the thesis for deposit in the UM repository.

We kindly request you submit the whole thesis as one document (i.e. not per chapter) and only once to prevent unnecessary draws on the maximum number of submissions, as our contract provides a limited number of checks.

iThenticate FAQ

similarity search phd thesis

Contact & Support

Ask your librarian - contact a library specialist.

Citation analysis of Ph.D. theses with data from Scopus and Google Books

  • Open access
  • Published: 24 October 2021
  • Volume 126 , pages 9431–9456, ( 2021 )

Cite this article

You have full access to this open access article

  • Paul Donner   ORCID: orcid.org/0000-0001-5737-8483 1  

5076 Accesses

3 Citations

9 Altmetric

Explore all metrics

This study investigates the potential of citation analysis of Ph.D. theses to obtain valid and useful early career performance indicators at the level of university departments. For German theses from 1996 to 2018 the suitability of citation data from Scopus and Google Books is studied and found to be sufficient to obtain quantitative estimates of early career researchers’ performance at departmental level in terms of scientific recognition and use of their dissertations as reflected in citations. Scopus and Google Books citations complement each other and have little overlap. Individual theses’ citation counts are much higher for those awarded a dissertation award than others. Departmental level estimates of citation impact agree reasonably well with panel committee peer review ratings of early career researcher support.

Similar content being viewed by others

Researchgate versus google scholar: which finds more early citations.

Mike Thelwall & Kayvan Kousha

similarity search phd thesis

The influence of discipline consistency between papers and published journals on citations: an analysis of Chinese papers in three social science disciplines

similarity search phd thesis

The Use of Google Scholar for Tenure and Promotion Decisions

Christopher R. Marsicano, John M. Braxton & Alexander R. K. Nichols

Avoid common mistakes on your manuscript.

Introduction

In this article we present a study on the feasibility of Ph.D. thesis citation analysis and its potential for studies of early career researchers (ECR) and for the rigorous evaluation of university departments. The context is the German national research system with its characteristics of a very high ratio of graduating Ph.D.’s to available open job positions in academia, a distinct national language publication tradition in the social sciences and humanities and slowly unfolding change from a traditional apprenticeship-type Ph.D. system to a grad school type system. The first nationwide census in Germany reported 152,300 registered active doctoral students in Germany (Vollmar 2019 ). In the same year, 28,404 doctoral students passed their exams in Germany (Statitisches Bundesamt 2018 ). Both universities and science and higher education policy attach high value to doctoral training and consider it a core task of the university system. For this reason, doctoral student performance also plays an important role in institutional assessment systems.

While there is currently no national scale research assessment implemented in Germany, all German federal states have introduced formula-based partial funding allocation systems for universities. In most of these, the number of Ph.D. candidates is a well-established indicator. Most universities also partially distribute funds internally by similar systems. Such implementations can be seen as incomplete as they do not take into account the actual research output of Ph.D. candidates. In this contribution we investigate if citation analysis of doctoral theses is feasible on a large scale and can conceptually and practically serve as a complement to current operationalizations of ECR performance. For this purpose we study the utility of two citation data sources, Scopus and Google Books. We analyze the obtained citation data at the level of university departments within disciplines.

Doctoral studies

The doctoral studies phase can theoretically be conceived as a status transition period. It comprises a status passage process from apprentice to formally acknowledged researcher and colleague in the social context of a scientific community (Laudel and Gläser 2008 ). Footnote 1 The published doctoral thesis and its public defense are manifest proof of the fulfilment of the degree criterion of independent scientific contribution, marking said transition. The scientific community, rather than the specific organization, collectively sets the goals and standards of work in the profession, and experienced members of a community of peers judge and grade the doctoral work upon completion. Footnote 2 Yet the specific organization also plays a very important role. The Ph.D. project and dissertation are closely associated with the hosting university as it is this organization that provides the environmental means to conduct the Ph.D. research, as a bare minimum the supervision by professors and experienced researchers, but often also formal employment with salary, workspace and facilities. And it is also the department ( Fakultät ) which formally confers the degree after passing the thesis review and defense.

As a rule, it is a formal requirement of doctoral studies that the Ph.D. candidates make substantial independent scientific contributions and publish the results. The Ph.D. thesis is a published scientific work and can be read and cited by other researchers. The extent to which other researchers make use of these results is reflected in citations to the work and is in principle amenable to bibliometric citation analysis (Kousha and Thelwall 2019 ). Citation impact of theses can be seen as a proxy of the recognition of the utility and relevance of the doctoral research results by other researchers. Theses are often not published in an established venue and are hence absent from the usual channels of communication of the research front, more so in journal-oriented fields, whereas in book-oriented fields, publication of theses through scholarly publishers is common. We address this challenge by investigating the presence of dissertation citations in data sources hitherto not sufficiently considered for this purpose in what follows.

Research contribution of early career researchers and performance evaluation in Germany

Almost all universities in Germany are predominantly tax-funded and the consumption of these public resources necessitates a certain degree of transparency to establish and maintain the perceived legitimacy of the higher education and research system. Consequently, universities and their subdivisions are increasingly subjected to evaluations. The pressure to participate in evaluation exercises, or in some cases the bureaucratic directive to do so by the responsible political actors, in turn, derives from demands of the public, which holds political actors accountable for the responsible spending of resources appropriated from net tax payers. Because the training of Ph.D. candidates is undisputedly a core task of universities, it is commonly implemented as an important component or dimension in university research evaluation.

While there is no official established national-scale research evaluation exercise in Germany (Hinze et al. 2019 ), the assessment of ECR performance plays a very importent role in evaluation and funding of universities and in the systems of performance-based funding within universities. In the following paragraphs we will shows this with several examples while critically discussing some inadequacies of the extant operationalizations of the ECR performance dimensions, thereby substantiating the case for more research into the affordance of Ph.D. thesis citation analysis.

The Council of Science and Humanities ( Wissenschaftsrat ) has conducted four pilot studies for national-scale evaluations of disciplines in universities and research institutes ( Forschungsrating ). While the exercises were utilized to test different modalities Footnote 3 , they all followed a basic template of informed peer review by appointed expert committees along a number of prescribed performance dimensions. The evaluation results did not have any serious funding allocation or restructuring consequences for the units. In all exercises, the dimension of support for early career researchers played a prominent role next to such dimensions as research quality, impact/effectivity, efficiency, and transfer of knowledge into society. Footnote 4 In all four exercises, the dimension was operationalized with a combination of quantitative and qualitative criteria.

As the designation ‘support for early career researchers’ suggests, the focus was primarily on the support structures and provisions that the assessed units offered, but the outcomes or successes of these support environments also played a role. Yet, some of the applied indicators are more in line with a construct such as the performance, or success, of the ECRs themselves, namely, first appointments of graduates to professorships, scholarships or fellowship of ECRs (if granted externally of the assessed unit), and awards. Footnote 5 As for the difference between the concept of the efforts expended for ECRs and the concept of the performance of ECRs, it appears to be implied that the efforts cause the performance, but this is far from self-evident. There may well be extensive support programs without realized benefits or ECRs achieving great success despite a lack of support structures. For this implied causal connection to be accepted, its mechanism should first be worked out and articulated and then be empirically validated, which was not the case in the Forschungsrating evaluation exercises. Footnote 6

No bibliometric data on Ph.D. theses was employed in the Forschungsrating exercises (Wissenschaftsrat 2007 , 2008 , 2011 , 2012 ). However, it stands to reason that citation analysis of theses might provide a valuable complementary tool if a more sound operationalization of the dimension of the performance of ECRs is to be established in similar future assessments. As for the publications of ECRs besides doctoral theses, these have been included in the other dimensions in which publications were used as criteria without special consideration. Footnote 7

There is a further area of university evaluation in which a performance indicator of ECRs, namely the absolute number of Ph.D. graduates over a specific time period, is an important component. At the time of writing, systems of partial funding allocation from ministries to states’ universities across all German federal states are well established. In these systems, universities within a state compete with one another for a modest part of the total budget based on fixed formulas relating performance to money. The performance based funding systems, different for each state, all include ‘research’ among their dimensions, and within it, the number of graduated Ph.D.’s is the second most important indicator after the acquired third party funding of universities (Wespel and Jaeger 2015 ). In direct consequence, similar systems have also found widespread application to distribute funds across departments within universities (Jaeger 2006 ; Niggemann 2020 ). These systems differ across universities. If only the number of completed Ph.D.’s is used as an indicator, then the quality of the research of the graduates does not matter in such systems. It is conceivable that graduating as many Ph.D.’s as possible becomes prioritized at the expense of the quality of ECR research and training.

A working group tasked by the Federal Ministry of Education and Research to work out an indicator model for monitoring the situation of early career researchers in Germany proposed to consider the citation impact of publications as an indicator of outcomes (Projektgruppe Indikatorenmodell 2014 ). Under the heading of “quality of Ph.D.—disciplinary acceptance and possibility of transfer” the authors acknowledge that, in principle, citation analysis of Ph.D. theses is possible, but citation counts do not directly measure scientific quality, but rather the level of response to, and reuse of, publications (impact). Moreover, it is stated that the literature of the social sciences and humanities are not covered well in citation indexes and theses are generally not indexed as primary documents (p. 136). Nevertheless, this approach is not to be rejected out of hand. Rather, it is recommended that the prospects of thesis citation analysis be empirically studied to judge its suitability (p. 137).

Another motivation for the present study was the finding of the National Report on Junior Scholars that even though “[j]unior scholars make a telling contribution to developing scientific and social insights and to innovation” (p. 3) the “contribution made by junior scholars to research and knowledge sharing is difficult to quantify in view of the available data” (Consortium for the National Report on Junior Scholars 2017 , p. 19).

To sum up, the foregoing discussion establishes (1) that there is a theoretically underdeveloped evaluation practice in the area of ECR support and performance, and (2) that a need for better early career researcher performance indicators on the institutional level has been suggested to science policy actors. This gives occasion to explore which, if any, contribution bibliometrics can make to a valid and practically useful assessment.

Prior research

Citation analysis of dissertation theses.

There are few publications on citation analysis of Ph.D. theses as the cited documents, as opposed to studies of the documents cited in theses, of which there are plenty. Yoels ( 1974 ) studied citations to dissertations in American journals in optics, political science (one journal each), and sociology (two journals) from the 1955 to 1969 volumes. In each case, several hundred citations in total to all Ph.D. theses combined were found, with a notable concentration on origins of Ph.D.’s in departments of high prestige – a possible first hint of differential research performance reflected in Ph.D. thesis citations. Non-US dissertations were cited only in optics. Author self-citations were very common, especially in optics and political science. While citations peaked in the periods of 1–2 or 3–5 years after the Ph.D. was awarded, they continued to be cited to some degree as much as 10 years later. According to Larivière et al. ( 2008 ), dissertations only account for a very small fraction of cited references in the Web of Science database. The impact of individual theses was not investigated. This study used a search approach in the cited references, based on keywords for theses and filtering, which may not be able to discover all dissertation citations. Kousha and Thelwall ( 2019 ) investigated Google Scholar citation counts and Mendeley reader counts for a large set of American dissertations from 2013 to 2017 sourced from ProQuest. This study did not take into account Google Books. Of these dissertations, 20% had one or more citations (2013: over 30%, 2017: over 5%) while 16% had at least one Mendeley reader. Average citation counts were comparatively high in the arts, social sciences, and humanities, and low in science, technology, and biomedical subjects. The authors evaluated the citation data quality and found that 97% of the citations of a sample of 646 were correct. As for the publication type of the citing documents, the majority were journal articles (56%), remarkably many were other dissertations (29%), and only 6% of citations originated from books. This suggests that Google Books might be a relevant citation data source instead of, or in addition to, Google Scholar.

More research has been conducted into the citation impact of thesis-related journal publications. Hay ( 1985 ) found that for the special case of a small sample from UK human geography research, papers based on Ph.D. thesis work accrued more citations than papers by established researchers. In a recent study of refereed journal publications based on US psychology Ph.D. theses, Evans et al. ( 2018 ) found that they were cited on average 16 times after 10 years. The citation impact of journal articles to which Ph.D. candidates contributed (but not of dissertations) has only been studied on a large scale for the Canadian province of Quèbec (Larivière 2012 ). The impact of journal papers with Ph.D. candidates’ contribution was contrasted to all other papers with Quèbec authors in the Web of Science database. As the impact of these papers, quantified as average of relative citations, was close to that of the comparison groups in three of four broad subject areas, it can be tentatively assumed that the impact of doctoral candidates’ papers was on par with that of their more experienced colleagues. The area with a notable difference between groups was arts and humanities, in which the coverage of publication output in the database was less comprehensive because a lot of research is published in monographs, and in which presumably many papers were written in French, another reason for lower coverage.

While these papers are not concerned with citations to dissertations, they do suggest that the research of Ph.D.’s is as impactful as that of other colleagues. To the best of our knowledge, no large scale study has been conducted on the citation impact of German theses on the level of individual works or on the level of university departments. We so far have scant information on the citation impact of dissertation theses, therefore the current study aims to fill this gap by a large scale investigation of citations received by German Ph.D. theses in Scopus and Google Books.

Causes for department-level performance differences

As we wish to investigate performance differences between departments of universities by discipline as reflected by thesis citations, we next consider the literature on plausible reasons for such performance differences which can result in differences in thesis citation impact. We do not consider individual level reasons for performance differences such as ability, intrinsic motivation, perseverance, and commitment.

One possible reason for cross-department performance differences is mutual selectivity of Ph.D. project applicants and Ph.D. project supervisors. In a situation in which there is some choice between the departments at which prospective Ph.D. candidates might register and some choice between the applicants a prospective supervisor might accept, out of self-interest both sides will seek to optimize their outcomes given their particular constraints. That is, applicants will opt for the most promising department for their future career while supervisors or selection committees, and thus departments, will attempt to select the most promising candidates, perhaps those who they judge most likely to contribute positively to their research agenda. Both sides can take into account a variety of criteria, such as departmental reputation or candidates’ prior performance. This is part of the normal, constant social process of mutual evaluation in science. However, in this case, the mutual evaluation does not take place between peers, that is, individuals of equal scientific social status. Rather, the situation is characterized by status inequality (superior-inferior, i.e. professor-applicant). Consequently, an applicant may well apply to her or his preferred department and supervisor, but the supervisor or the selection committee makes the acceptance decision. In practice however, there are many constraints on such situations. For example, as described above, the current evaluation regime rewards the sheer quantity of completed Ph.D.’s.

Once the choices are made, Ph.D. candidates at different departments can face quite different environments, more or less conducive to research performance (which, as far as they were aware of them and were able to judge them, they would have taken into consideration, as mentioned). For instance, some departments might have access to important equipment and resources, others not. There may prevail different local practices in time available for the Ph.D. project for employed candidates as opposed to expected participation in groups’ research, teaching, and other duties (Hesli and Lee 2011 ).

Ph.D. candidates may benefit from the support, experience and stimulation of the presence of highly accomplished supervisors. Experienced and engaged supervisors teach explicit and tacit knowledge and can serve as role models. Long and McGinnis ( 1985 ) found that the performance of mentors was associated with Ph.D.’s publication and citation counts. In particular, citations were predicted by collaborating with the mentor and the mentor’s own prior citation counts. Mentors’ eminence only had a weak positive effect on the publication output of Ph.D.’s who actively collaborated with them. Similarly, Hilmer and Hilmer ( 2007 ) report that advisors’ publication productivity is associated with candidate’s publication count. However, there are multiple professors or other supervisors at any department, which causes variation within departments if the department and not the supervisor is used as a predictive variable. Between departments it is then the concentration of highly accomplished supervisors that may cause differences. Beyond immediate supervisors, a more or less supportive research environment can offer opportunities for learning, cooperation or access to personal networks. For example, Kim and Karau ( 2009 ) found that support from faculty, through the development of research skills, lead to higher publication productivity of management Ph.D. candidates. Local work culture and local expectations of performance may elicit behavioral adjustment (Allison and Long 1990 ).

In summary, prior research shows that there are several reasons to expect department-level differences of Ph.D. research quality (and its reproduction and reinforcement) which might be reflected in thesis citation impact. But it needs to be noted that the present study cannot serve towards shedding light on which particular factors are associated with Ph.D. performance in terms of citation impact. It is limited to testing if there are any department-level differences on this measure.

Citation counts and scientific impact of dissertation theses

We have argued above that citation analysis of theses could be a complementary tool for quantitative assessment of university departments in terms of the research performance of early career researchers. Hence it needs to be established that citation counts of dissertations are in fact associated with a conception of the impact of research.

As outlined by Hemlin ( 1996 ), “[t]he idea [of citation analysis] is that the more cited an author or a paper is by others, the more attention it has received. This attention is interpreted as an indicator of the importance, the visibility, or the impact of the researcher or the paper in the scientific community. Whether citation measures also express research quality is a highly debated issue.” Hemlin reviewed a number of studies of the relationship between citations and research quality but was not able to make a definite conclusion: “it is possible that citation analysis is an indicator of scientific recognition, usefulness and, to some unknown extent, quality.” Researchers cite for a variety of reasons, not only or primarily to indicate the quality of the cited work (Aksnes et al. 2019 ; Bornmann and Daniel 2008 ). Nevertheless, work that is cited usually has some importance for the citing work. Even citations classified in citation behavior studies as ‘perfunctory’ or ‘persuasive’ are not made randomly. On the contrary, for a citation to persuade anyone, the content of the cited work needs to be convincing rather than ephemeral, irrelevant, or immaterial. Citation counts are thus a direct measure of the utility, influence, and importance of publications for further research (Martin and Irvine 1983 , sec. 6). Therefore, as a measure of scientific impact, citation counts have face validity. They are a measure of the concept itself, though a noisy one. Not so for research quality.

Highly relevant for the topic of the present study are the early citation impact validation studies by Nederhof and van Raan ( 1987 ), Nederhof and van Raan ( 1989 ). These studied the differences in citation impact of publications produced during doctoral studies of physics and chemistry Ph.D. holders, comparing those awarded the distinction ‘cum laude’ for their dissertation based on the quality of the research with other graduates without this distinction (cum laude: 12% of n = 237 in chemistry, 13% of n = 138 in physics). In physics, “[c]ompared to non-cumlaudes, cumlaudes received more than twice as many citations overall for their publications, which were all given by scientists outside their alma mater” (Nederhof and van Raan 1987 , p. 346). In fact, differences in citation impact of papers between the groups are already apparent before graduation, that is, before the conferral of the cum laude distinction on the basis of the dissertation. And interestingly, “[a]fter graduation, citation rates of cumlaudes even decline to the level of non-cumlaudes” (p. 347) leading the authors to suggest that “the quality of the research project, and not the quality of the particular graduate is the most important determinant of both productivity and impact figures. A possible scenario would be that some PhD graduates are choosen carefully by their mentors to do research in one of the usually rare very promising, interesting and hot research topics currently available. Most others are engaged in relatively less interesting and promising graduate research projects” (p. 348). The results in chemistry are very similar: “Large difference in impact and productivity favor cumlaudes three to 2 years before graduation, differences which decrease in the following years, although remaining significant. [...] Various sceptics have claimed that bibliometric measures based on citations are generally invalid. The present data do not offer any support for this stance. Highly significant differences in impact and productivity were obtained between two groups distinguished on a measure of scientific quality based on peer review (the cum laude award)” (Nederhof and van Raan 1989 , p. 434).

In Germany, a system of four passing marks and one failing mark is commonly used. The better the referees judge the thesis, the higher the mark. Studies investigating the association of level of mark and citation impact of theses or thesis-associated publications are as of yet lacking. The closest are studies on medical doctoral theses from Charité. Oestmann et al. ( 2015 ) provide a correlational study of medical doctoral degree marks (averages of thesis and oral exams marks) and the publications associated with the theses from one institution, Charité University Medicine Berlin. Their data for 1992–2014 shows a longitudinal decrease of the incidence of the third best mark and an increase of the second best mark. For samples from 3 years (1998, 2004, 2008) for which publication data were collected, an association between the level of the mark and the publication productivity was detected. Both the chance to publish any peer-reviewed articles and the number of articles increase with the level of the mark. The study was extended in Chuadja ( 2021 ) with publication data for 2015 graduates. It was found that the time to graduation covaries with the level of the mark. For 2015 graduates, the average 5 year Journal Impact Factors for thesis-associated publication increase with the level of the graduation mark in the sense that theses awarded better marks produced publications in journals with higher Impact Factors. As little as these findings say about the real association of thesis research quality and citation impact, they suggest enough to motivate more research into this relationship.

Research questions

The following research questions will be addressed:

How often are individual Ph.D. theses cited in the journal and book literature?

Does Google Books contain sufficient additional citation data to warrant its inclusion as an additional data source alongside established data sources?

Can differences between universities within a discipline explain some of the variability in citation counts?

Are there noteworthy differences in Ph.D. thesis citation impact on the institutional level within disciplines?

Are the citation counts of Ph.D. theses associated with their scientific quality?

To test whether or not dissertation citation impact is a suitable indicator of departmental Ph.D. performance, citation data for theses needs to be collected, aggregated and studied for associations with other relevant indicators, such as doctorate conferrals, drop-out rates, graduate employability, thesis awards, or subjective program appraisals of graduates. As a first step towards a better understanding of Ph.D. performance, we conducted a study on citation sources for dissertations. The present study is restricted to monograph form dissertations. These also include monographs that are based on material published as articles. However, to be able to assess the complete scientific impact of a Ph.D. project it is necessary to also include the impact of papers which are produced in the context of the Ph.D. project, for both cumulative publication-based theses and for theses only published in monograph form. Because of this, the later results should be interpreted with due caution as we do not claim completeness of data.

Dissertations’ bibliographical data

There is presently no central integrated source for data on dissertations from Germany. The best available source is the catalog of the German National Library (Deutsche Nationalbibliothek, DNB). The DNB has a mandate to collect all publications originating from Germany, including Ph.D. theses. This source of dissertation data has been found useful for science studies research previously (Heinisch and Buenstorf 2018 ; Heinisch et al. 2020 ). We downloaded records for all Ph.D. dissertations from the German National Library online catalog in April 2019 using a search restriction in the university publications field of “diss*”, as recommended by the catalog usage instructions, and publication year range 1996–2018. Records were downloaded by subject fields in the CSV format option. Footnote 8 In this first step 534,925 records were obtained. In a second step, the author name and work title field were cleaned and the university information extracted and normalized and non-German university records excluded. We also excluded records assigned to medicine as a first subject class which were downloaded because they were assigned to other classes as well. As the dataset often contained more than one version of a particular thesis because different formats and editions were cataloged, these were carefully de-duplicated. In this process, as far as possible the records containing the most complete data and describing the temporally earliest version were retained as the primary records. Variant records were also kept in order to later be able to collect citations to all variants. This reduced the dataset to a size of 361,971 records. Of these, about 16% did not contain information on the degree-granting university. As the National Library’s subject classification system was changed during the covered period (in 2004), the class designations were unified based on the Library’s mapping and aggregated into a simplified 40-class system. Footnote 9 If more than one subject class was assigned, only the first was retained.

Citation data

Citation data from periodicals was obtained from a snapshot of Scopus data from May 2018. Scopus was chosen over Web of Science as a source of citation data because full cited titles for items not contained as primary documents in Web of Science have only recently been indexed. Before this change, titles were abbreviated so inconsistently and to such short strings as to be unusable, while this data is always available in unchanged form in Scopus if it is available in the original reference. Cited references data was restricted to non-source citations, that is, references not matched with Scopus-indexed records. Dissertation bibliographical information (author name, publication year and title) for primary and secondary records was compared to reference data. If the author family name and given name first initial matched exactly and the cited publication year was within ± 1 year of the dissertation publication year, then the title information was further compared as follows. The dissertation’s title was compared to both Scopus cited reference title and cited source title, as we found these two data fields were both used for the cited thesis title. Before comparison, both titles were truncated to the character length of the shorter title. If the edit distance similarity between titles was greater than 75 out of 100, the citation was accepted as valid and stored. We furthermore considered the case that theses might occasionally be indexed as Scopus source publications. We used the same matching approach as outlined above to obtain matching Scopus records restricted to book publication types. This resulted in 659 matched theses. In addition, matching by ISBN to Scopus source publications resulted in 229 matched theses of which 50 were not matched in the preceding step. The citations to these matched source publications were added to the reference matched citations after removing duplicates. Citations were summed across all variant records while filtering out duplicate citations.

In addition, we investigated the utility of Google Books as a citation data source. This is motivated by the fact that many Ph.D. theses are published in the German language and in disciplines favoring publication in books over journal articles. Citation search in Google Books has been made accessible to researchers by the Webometric Analyst tool, which allows for largely automated querying with given input data (Kousha and Thelwall 2015 ). We used Webometric Analyst version 2.0 in April and May 2019 to obtain citation data for all Ph.D. thesis records. We only used primary records, not variants, as the collection process takes quite a long time. Search was done with the software’s standard settings using author family name, publication year and six title words and subsequent result filtering was employed with matching individual title words rather than exact full match. We additionally removed citations from a number of national and disciplinary bibliographies and annual reports because these are not research publications but lists of all publications in a discipline or produced by a certain research unit. We also removed Google Books citations from sources that were indexed in Scopus (10,958 citations) as these would otherwise be duplicates.

Google Scholar was not used as a citation data source, because it includes a lot of grey literature and there is no possibility to restrict citation counts to citations from journal and book literature. It has alarming rates of incorrect citations (García-Pérez 2010 , p. 2075), however Kousha & Thelwall (2015, p. 479) found the citation error rate for Ph.D. theses in Google Scholar to be quite low.

Dissertation award data

For later validation purposes we collected data on German dissertation awards from the web. We considered awards for specific disciplines granted in a competitive manner based on research quality by scholarly societies, foundations and companies. A web search was conducted for awards, either specifically for dissertations or for contributions by early career researchers which mention dissertations besides other works. Only awards granted by committees of discipline experts and awarded by Germany-based societies etc. were considered, we did not include internationally oriented awards. In general, these awards are given on the basis of scientific quality and only works published in the preceding one or 2 years are accepted for submission. We were able to identify 946 Ph.D. theses that received one or more dissertation awards from a total of 122 different awards. More details can be found in the “Appendix”, Table 5 .

A typical example is the Wilhelm Pfeffer Prize of the German Botanical Society, which is described as follows: “The Wilhelm Pfeffer Prize is awarded by the DBG’s Wilhelm Pfeffer Foundation for an outstanding PhD thesis (dissertation) in the field of plant sciences and scientific botany.” Footnote 10 The winning thesis is selected by the board of the Foundation. If no work achieves the scientific excellence expected by the board, no prize is awarded. Footnote 11

Citation data for Ph.D. theses

In total we were able to obtain 329,236 Scopus citations and 476,495 Google Books citations for the 361,971 Ph.D. thesis records. There was an overlap of about 11,000 citations from journals indexed in both sources, which was removed from the Google Books data. The large majority of Scopus citations was found with the primary thesis records only (95%). Secondary (variant) thesis records and thesis records matched as Scopus source document records contributed 5% of the Scopus citations. Scopus and Google Books citation counts are modestly correlated with Pearson’s r = 0.20 ( \(p < 0.01\) , 95% CI 0.197–0.203). Table 1 gives an overview of the numbers of citations for theses published in different years. Footnote 12 One can observe a minor overall growth in annual dissertations and modest peak in citations in the early years of the observation period. Overall, and consistently for all thesis publication years, theses are cited more often in Google Books than in Scopus by a ratio of about 3 to 2. Hence, in general, German Ph.D. theses were cited more often in the book literature covered by Google Books than in the periodical literature covered by Scopus. The average number of citations per thesis seems to stabilize at around 3.5, as the values for 1996–2003 are all of that magnitude. As citations were collected in 2019, these works needed about 15 years to reach a level at which they did not increase further. Whereas Kousha and Thelwall ( 2019 ) found that 20% of the American dissertations from the period 2013–2017 were cited at least once in Google Scholar, the corresponding figure from our data set is 30% for the combined figure of both citation data sources.

We studied author self-citations of dissertations, both by comparing the thesis author to all publication authors of a citing paper (all-authors self-citations) and only to the first author. We used exact match or Jaro-Winkler similarity greater than 90 out of 100 for author name comparison (Donner 2016 ). Considering only the first authors of publications is more lenient in that it does not punish an author for self-citation that could possibly be suggested by co-authors and is at least endorsed by them (Glänzel and Thijs 2004 ). For the Google Books citation corpus we find only 8366 all-authors self-citations (1.7%) and 5711 first author self-citations (1.2%). Footnote 13 In the Scopus citations there are 52,032 all-author self-citations (15.6%) and 31,260 first-author self-citations (9.4%). Overall this amounts to an all-author self-citation rate of 7.5% and a first-author self-citation rate of 4.6%, quite lower rates than Yoels ( 1974 ). We do not exclude any self-citations in the following analyses.

Figure  1 displays the citation count distribution for theses published before 2004, in which citation counts are likely to not increase any further. In this subset, 58% of dissertations have one or more citations. Ph.D. theses exhibit the typical highly skewed and long-tailed citation count distribution.

figure 1

Citation count distribution for German Ph.D. theses from 1997–2003 (values greater than 10 citations not shown), n = 118,447

The distributions of theses and citations by data source over subject classes are displayed in Table 2 . There are striking differences in the origin of the major part of citations across disciplines. Social sciences and humanities such as education, economics, German language and literature, history, and especially law are cited much more often in Google Books. The opposite holds in natural science subjects like biology or physics, and in computer science and mathematics, where most citations come from Scopus. This table also indicates that the overall most highly cited dissertations can be found in the humanities (history, archeology and prehistory, religion) and that natural science dissertations are poorly cited (biology, physics and astronomy, chemistry). Much of this difference is probably because in the latter subjects, Ph.D. project results are almost always communicated in journal articles first and citing a dissertation is rarely necessary.

Validation of citation count data with award data

In order to judge whether thesis citation counts can be considered to include a valid signal of scientific quality (research question 5) we studied the citation impact of theses that received dissertation awards compared to those which did not. High citation counts, however, can not simply be equated with high scientific quality. As a rule, the awards for which we collected data are conferred by committees of subject experts explicitly on the basis of scientific quality and importance. But if the content of theses has been published in journal articles before it was published in a thesis it is possible that awards juries might have been influenced by citation counts of these articles.

Comparing the 946 dissertations that were identified as having received a scientific award directly with the non-awarded dissertations we find that the former received on average 3.9 citations while the latter received on average 2.2 citations. To factor out possible differences across subjects and time we match each award-winning thesis with all non-awarded theses of the same subject and year, and calculate the average citation count of this comparable publication set. Award theses obtain 3.9 citations and matched non-award theses 1.7 citations. This shows that Ph.D. theses that receive awards on average are cited more often than comparable theses and indicates that citation counts partially indicate scientific quality as sought out by award committees. Nevertheless, not every award-winning thesis need be highly cited nor every highly cited thesis be of high research quality and awarded a prize. The differences reported here hold on average for large numbers of observations, but do not imply a strong association between scientific quality and citation count on the level of individual dissertations. This is important to note lest the false conclusion be drawn that merely because a thesis is not cited or rarely cited, it is of low quality. Such a view must be emphatically rejected as it is not supported by the data.

A possible objection to the use of award data for validating the relationship of citation counts and the importance and quality of research is that it might be the signal of the award itself which makes the publications more visible to potential citers. In other words, a thesis is highlighted and brought to the attention of researchers by getting an award and high citation counts are a result not of the intrinsic merit of the reported research but merely of the raised visibility. The study by Diekmann et al. ( 2012 ) has scrutinized this hypothesis. They studied citation counts (Social Sciences Citation Index) of 102 papers awarded the Thyssen award for German social science journal papers and a random control sample of other publications from the same journals. The award winners are determined by a jury of experts and there are first, second, and third rank awards each year. It was found that awarded papers were cited on average six times after 10 years, control articles two times. Moreover, first rank award articles were cited 9 times, second rank articles 6 times, and third rank articles 4 times on average. The jury decides in the year after the publication of the articles. The authors argue that publications citing awarded articles in the first year after its publication can not possibly have been influenced by the award. For citation counts of a 1-year citation window, awarded articles are cited 0.8 times, control group articles 0.2 times on average. And again, the ranks of the awards correspond to different citation levels. Thus it is evident that the citation reception of the articles are different even before the awards have been decided. Citing researchers and expert committee independently agree on the importance of these articles.

We can replicate this test with our data by restricting the citations to those received in the year of publication of the thesis and the next year. This results in average citation counts of 0.040 for theses which received awards and 0.012 for theses which did not receive any award. Award-winning Ph.D. theses are cited more than three times as often as other theses even before the awards have been granted and before any possible publicity had enough time to manifest itself in increased citation counts.

Application: a preliminary study of citation impact differences between departments

In this section we consider research questions 3 and 4, which are concerned with the existence of differences in Ph.D. thesis citation impact between universities in the same discipline and their magnitude. This application is preliminary because only the citation impact of the thesis but not of the thesis-related publications is considered here and because we use a subject classification based on that of the National Library. Footnote 14 In order to mitigate against these limitations as much as possible, we study here only two subjects from the humanities (English and American language and literature, henceforth EALL) and the social sciences (sociology) which are characterized by Ph.D. theses mainly published as monographs rather than as articles and thus show relatively high dissertation citation counts. These are also disciplines specifically identifiable in the classification system and as distinct university departments. We furthermore restrict the thesis publication years to the years covered by the national scale pilot research assessment exercises discussed in the introduction section which were carried out by the German Council of Science and Humanities in these two disciplines (Wissenschaftsrat 2008 , 2012 ) in order to be able to compare the results and to test if the number of observations in typical evaluation periods yield sufficient data for useful results.

We use multilevel Bayesian negative binomial regressions in which the observations (theses) are nested in the grouping factor universities, estimated with the R package brms , version 2.8.0 (Bürkner 2017 ). By using a multilevel model we can explicitly model the correlation of observations within a group, that is to say, take into account the characteristics of the departments which affect all Ph.D. candidates of a department and their research quality. The negative binomial response distribution is appropriate for the characteristic highly skewed, non-negative citation count distribution. The default prior distributions of brms were used. Model estimations are run with two MCMC chains with 2500 iterations each, of which the first 500 are for sampler warmup. As population level variables we include a thesis language dummy (German [reference category], English, Unknown/Other), the publication year, and dummies for whether the dissertation received an award (reference category: no award received). There were no awards identified for EALL in the observation period, so this variable only applies to the Sociology models. For the language variable we used the data from the National Library where available and supplemented it with automatically identified language of the dissertation title using R package cld3 , version 1.1. Languages other than English and German and theses with unidentifiable titles were coded as Unknown/Other.

Models are run for the two disciplines separately, once without including the university variable (null models) and once including it as a random intercept, making these multilevel models. If the multilevel model shows a better fit, i.e. can explain the data better, this would indicate significant variation between the different university departments in a discipline and higher similarity of citation impact of theses within a university than expected at random. In other words, the null model assumes independence of observations while the multilevel model controls for correlation of units within a group, here the citation counts of theses from a university. The results are presented in Table 3 . The coefficient of determination (row ‘R 2 ’) is calculated according to Nakagawa et al. ( 2017 ).

Regarding the population level variables, it can be seen that the publication year has a negative effect in EALL (younger theses received less citations, as expected) but no significant effect in Sociology. As the Sociology models data is from an earlier time period, this supports the notion that the citation counts used in the Sociology models have stabilized but those from EALL have not. According to the credible intervals, while there is no significant language effect on citations in the Sociology null model, controlling for the grouping structure reveals the language to be significant predictor. In EALL, English language theses received substantially more citations than German language theses. There is a strong positive effect in Sociology from having received an award for the Ph.D. thesis. If we compare the null models with their respective multilevel models, that is A to B and C to D, we can see that introducing the grouping structure does not affect the population level variable effects other than language for Sociology. In both disciplines, the group effect (standard deviation of the random intercept) is significantly above zero and the model fit in terms of R 2 improved, indicating that the hierarchical model is more appropriate and that the university department is a significant predictor. However, the values of the coefficients of determination are small, which suggests that it is not so much the included population level predictors and the group membership, but additional unobserved thesis-level characteristics that affect citation count. In addition, this means it is not possible to estimate with any accuracy a particular thesis’ citation impact only from knowing the department at which it originated. The estimated group effects describe the citation impact of particular university departments. The distribution of the estimates from models B and D with associated 95% credible intervals are displayed in Fig.  2 . It is evident that while there are substantial differences in the estimates as a whole, there is also large uncertainty about the exact magnitude about all effects, indicated by the wide credible intervals. This a consequence of the facts that, first, most departments produce theses across the range of citation impact and small differences in the ratios of high, middle and low impact theses determine the estimated group effects and, second, given the high within-group variability, there are likely too few observations in the applied time period to arrive at more precise estimates.

figure 2

Ph.D. thesis citation impact regression group effect estimates for a 49 sociology departments (2001–2005) and b 52 English and American language and literature departments (2004–2010). Means and 95% posterior probability ranges

The results of the above mentioned research evaluation of sociology departments in the Forschungsrating series of national scale pilot assessments allow for a comparison between the university group effects of Ph.D. thesis citation impact obtained in the present study and qualitative ordinal ratings given by the expert committee in the category ‘support of early career researchers.’ Footnote 15 In the sociology assessment exercise, the ECR support dimension was explicitly intended to reflect both the supportive actions of the departments and their successes. The reviewer panel put special weight on the presence of structured degree programs and scholarships and obtained professorship positions of graduates. Further indicators that were taken into account were the number of conferred doctorates, the list of Ph.D. theses with publisher information, and a self-report of actions and successes. This dimension was rated by the committee on a five point scale ranging from 1, ‘not satisfactory’, to 5, ‘excellent’ (Wissenschaftsrat 2008 , p. 22).

For 47 universities both an estimated citation impact score (group effect coefficient from the above regression) and an ECR support rating were available for sociology in the same observation period. A tabulated comparison is presented in Table 4 . The Kendall \(\tau\) rank correlation between these two variables at the department level is 0.36 ( \(p < 0.01\) ), indicating a moderate association, but the mean citation scores in the table do not exhibit a clear pattern of increase with increasing expert-rated ECR support. The bulk of departments were rated in the lower middle and middle categories, that is to say, the ratings are highly concentrated in this range, making distinctions quite difficult.

figure 3

ECR support expert committee rating and mean estimates of Ph.D. thesis citation impact of 47 German sociology departments (2001–2005)

This relationship is displayed in Fig.  3 . It can be seen that while the five rating groups do have some association with the citation analysis estimates, there is large variability within the groups, especially for categories 3 and 4. In fact, there is much overlap across all the rating groups in terms of citation impact scores. The university department with the highest citation impact effect estimate was rated as belonging to the middle groups of ECR support. In summary, the association between rated ECR support of a department and the impact of the department’s citation is demonstrable but moderate in size.

In this study we have demonstrated the utility of the combination of citation data from two distinct sources, Scopus and Google Books, for citation analysis of Ph.D. theses in the context of research evaluation of university departments. Our study has a number of limitations hampering its immediate implementation into practice. We did not have verified data about the theses produced by departments and used publication sets approximated by using a subject classification. We did not take into account the publications of Ph.D.’s other than the theses, such as journal articles, proceedings papers, and book chapters, which resulted in low citation counts in the sciences. These limitations must be resolved before any application for research evaluation and the present study is to be understood as a feasibility study only.

We now turn to the discussion of the results as they relate to the guiding research questions. The first research question concerned the typical citation counts of Ph.D. theses. We found that German Ph.D. theses were cited more often in the book literature covered by Google Books than in the periodical literature covered by Scopus and that it takes 10–15 years for citation counts to reach a stable level where they do not increase any further. At this stage, about 40% of theses remained uncited. We further found large differences in typical citation rates across fields. Theses are cited more often in the social sciences and humanities, especially in history and in archeology and prehistory. But citation rates were very low in physics and astronomy, chemistry and veterinary medicine. Furthermore, there were distinctive patterns of the origin of the bulk of citations between the two data sources, in line with the typical publication conventions of the sciences and the social sciences and humanities. Science theses received more citations from Scopus, that is, primarily the journal literature, than from the book literature covered by Google Books. The social sciences and humanities, on the other hand, obtained far more citations from the book literature covered in Google Books. Nevertheless, these fields’ theses do also receive substantial numbers of citations from the journal literature which must not be neglected. Thus with regard to our second research question we can state that Google Books is clearly a very useful citation data source for social sciences and humanities dissertations and most likely also for other publications beyond dissertations from these areas of research. Citations from Google Books are complementary to those from Scopus as they clearly cover a different literature; the two data sources have little overlap.

Our results of multilevel regressions allow us to affirm that there are clearly observable thesis citation impact differences between departments in a given discipline (research question 3). However, they are of small to moderate magnitude and the major part of citation count variation is found at the individual level (research question 4). Our results do not allow any statements about what factors caused these differences. They could be the result of mutual selection processes or of different levels of support or research capacity across departments. Our results of citation impact of departments, interpretable as collective performance of early career researchers at the university department level, are roughly in line with qualitative ratings of an expert committee in our case study of sociology. This association also does not rule out or confirm any possible explanations, as both department supportive actions and individual ECR successes were compounded in the rated dimension which furthermore showed only limited variation.

Research question 5 concerned the association of citation counts of Ph.D. theses and their research quality. Using information about dissertation awards we showed that theses which received such awards also received higher citation rates vis-à-vis comparable theses which did not win such awards. This result strongly suggests that if researchers cite Ph.D. theses, they do tend to prefer the higher quality ones, as recognized by award granting bodies and committees.

As a case in point for the status change approach we note that, in Germany, holding a Ph.D. degree is a requirement for applying for funding at the national public science funding body Deutsche Forschungsgemeinschaft.

While in other countries it is the norm that doctoral theses are evaluated by experts external to the university, this has traditionally not been the case in Germany (Wissenschaftsrat 2002 ). In fact, grading of the thesis by the Ph.D. supervisor is not considered inappropriate by the German Rectors’ Conference (Hochschulrektorenkonferenz, 2012 , p. 7).

The assessed disciplines were chemistry, sociology, electrical and information engineering, and English and American language and literature. Possibly the most important experimentally varied factor was the criterion by which research outputs were accredited to units: either all outputs in the reference period by researchers employed at the unit at assessment time (“current potential” principle) or all outputs of the units' research staff in the period, regardless of employment at assessment time (“work-done-at” principle).

As an exception, the committee on the evaluation of English and American language and literature considered the support of ECRs as a sub-dimension of the ‘enabling of research’ dimension, alongside third-party funding activity and infrastructures, and networks. However, within this category, it was given special importance.

Other applied indicators are extra-scientific in that they are indicators of compliance to science-external political directions, such as the share of female doctoral candidates.

Another issue deserving more scrutiny are the incentive structures promoted by the indicators. Indicators such as the number of granted Ph.D.’s and number of current Ph.D. candidates, which were applied in all exercises, could further exacerbate the Ph.D. oversupply in academia (Rogge and Tesch 2016 ).

It would be justified to (also) include publications by ECR other than theses in an ECR performance dimension, ideally appropriately weighted for ECR’s co-authorship contribution (Donner 2020 ).

The field of medicine was not included, because medical theses (for a Dr. med. degree) typically have lower research requirements and are therefore generally not commensurable to theses in other subjects (Wissenschaftsrat 2004 , pp. 74–75). It was not possible to distinguish between Dr. med. theses and other degree theses within the medicine category, which means that regular natural science dissertations on medical subjects are not included in this dataset if medicine was the first assigned class.

Classification mapping according to https://wiki.dnb.de/download/attachments/141265749/ddcSachgruppenDNBKonkordanzNeuAlt.pdf accessed 07/18/2019. It should be noted that some classes were not used for some time, for example electrical engineering was not used between 2004 and 2010 but grouped under engineering/mechanical engineering alongside mining/construction technology/environmental technology.

https://www.deutsche-botanische-gesellschaft.de/en/about-us/promoting-young-scientists/wilhelm-pfeffer-prize accessed 26/04/2021.

https://www.deutsche-botanische-gesellschaft.de/ueber-die-dbg/nachwuchsfoerderung/wilhelm-pfeffer-preis/satzung-pfeffer-stiftung §4, accessed 26/04/2021.

These figures are lower than the official numbers for completed doctoral exams as published by the Federal Statistical Office in Series 11/4/2 because medicine is not included.

As this is a very low figure, we manually checked the results and did not notice any issues. We note that about 58,000 (about 17%) of the Google Books citation records did not have any names of authors.

This is not ideal because its classes are not aligned with the delimitations of university departments. The National Library data do not contain information about the university department which accepted a thesis. As this exercise is not intended to be understood as an evaluation of university departments but only as a feasibility study, we still use the classification in lack of a better alternative. It goes without saying that for any actual evaluation use this course of action is not appropriate. Instead, a verified list of the dissertations of each unit would be required and all publications related to the theses would have to be included.

This is not possible for EALL because early career researcher support was not assessed separately but compounded with acquisition of third party funding and ‘infrastructures and networks’ into a dimension called ‘enabling of research’ and the ratings in this category were further differentiated by four sub-fields (Wissenschaftsrat 2012 , p. 19).

Aksnes, D. W., Langfeldt, L., & Wouters, P. (2019). Citations, citation indicators, and research quality: An overview of basic concepts and theories. Sage Open, 9 (1). https://doi.org/10.1177/2158244019829575 .

Allison, P. D., & Long, J. S. (1990). Departmental effects on scientific productivity. American Sociological Review, 55(4), 469–478. https://doi.org/10.2307/2095801 .

Article   Google Scholar  

Bornmann, L., & Daniel, H.-D. (2008). What do citation counts measure? A review of studies on citing behavior. Journal of Documentation, 64 (1), 45–80. https://doi.org/10.1108/00220410810844150 .

Bürkner, P.-C. (2017). brms: An R package for Bayesian multilevel models using Stan. Journal of Statistical Software, 80 (1), 1–28.

Chuadja, M. (2021). Promotionen an der Charité Berlin von 1998 bis 2015. Qualität, Dauer, Promotionstyp (Ph.D. thesis). Charité - Universitätsmedizin Berlin.

Consortium for the National Report on Junior Scholars. (2017). 2017 National Report on Junior Scholars. Statistical Data and Research Findings on Doctoral Students and Doctorate Holders in Germany. Overview of Key Results . Retrieved from https://www.buwin.de/dateien/buwin-2017-keyresults.pdf .

Diekmann, A., Näf, M., & Schubiger, M. (2012). Die Rezeption (Thyssen-) preisgekrönter Artikel in der “Scientific Community.” Kölner Zeitschrift für Soziologie und Sozialpsychologie, 64 (3), 563–581. https://doi.org/10.1007/s11577-012-0175-4 .

Donner, P. (2016). Enhanced self-citation detection by fuzzy author name matching and complementary error estimates. Journal of the Association for Information Science and Technology, 67 (3), 662–670. https://doi.org/10.1002/asi.23399 .

Donner, P. (2020). A validation of coauthorship credit models with empirical data from the contributions of PhD candidates. Quantitative Science Studies, 1 (2), 551–564. https://doi.org/10.1162/qss_a_00048 .

Evans, S. C., Amaro, C. M., Herbert, R., Blossom, J. B., & Roberts, M. C. (2018). “Are you gonna publish that?” Peer-reviewed publication outcomes of doctoral dissertations in psychology. PloS ONE, 13 (2), e0192219. https://doi.org/10.1371/journal.pone.0192219

García-Pérez, M. A. (2010). Accuracy and completeness of publication and citation records in the Web of Science, PsycINFO, and Google Scholar: A case study for the computation of h indices in Psychology. Journal of the American Society for Information Science and Technology, 61 (10), 2070–2085. https://doi.org/10.1002/asi.21372 .

Glänzel, W., & Thijs, B. (2004). Does co-authorship inflate the share of self-citations? Scientometrics, 61 (3), 395–404. https://doi.org/10.1023/b:scie.0000045117.13348.b1 .

Hay, A. (1985). Some differences in citation between articles based on thesis work and those written by established researchers: Human geography in the UK 1974–1984. Social Science Information Studies, 5 (2), 81–85. https://doi.org/10.1016/0143-6236(85)90017-1 .

Heinisch, D. P., & Buenstorf, G. (2018). The next generation (plus one): An analysis of doctoral student’s academic fecundity based on a novel approach to advisor identification. Scientometrics, 117 (1), 351–380. https://doi.org/10.1007/s11192-018-2840-5

Heinisch, D. P., Koenig, J., & Otto, A. (2020). A supervised machine learning approach to trace doctorate recipient’s employment trajectories. Quantitative Science Studies, 1 (1), 94–116. https://doi.org/10.1162/qss_a_00001

Hemlin, S. (1996). Research on research evaluation. Social Epistemology, 10 (2), 209–250. https://doi.org/10.1080/02691729608578815 .

Hesli, V. L., & Lee, J. M. (2011). Faculty research productivity: Why do some of our colleagues publish more than others? PS: Political Science & Politics, 44 (2), 393–408. https://doi.org/10.1017/S1049096511000242 .

Hilmer, C. E., & Hilmer, M. J. (2007). On the relationship between the student-advisor match and early career research productivity for agricultural and resource economics Ph.Ds. American Journal of Agricultural Economics, 89 (1), 162–175. https://doi.org/10.1111/j.1467-8276.2007.00970.x .

Hinze, S., Butler, L., Donner, P., & McAllister, I. (2019). Different processes, similar results? A comparison of performance assessment in three countries. In W. Glänzel, H. F. Moed, U. Schmoch, & M. Thelwall (Eds.), Springer handbook of science and technology indicators (pp. 465–484). Berlin: Springer. https://doi.org/10.1007/978-3-030-02511-3_18 .

Hochschulrektorenkonferenz. (2012). Zur Qualitätssicherung in Promotionsverfahren . Retrieved from https://www.hrk.de/positionen/beschluss/detail/zur-qualitaetssicherung-in-promotionsverfahren/ .

Jaeger, M. (2006). Leistungsbezogene Budgetierung an deutschen Universitäten - Umsetzung und Perspektiven. Wissenschaftsmanagement, 12 (3), 32–38.

Google Scholar  

Kim, K., & Karau, S. J. (2009). Working environment and the research productivity of doctoral students in management. Journal of Education for Business, 85 (2), 101–106. https://doi.org/10.1080/08832320903258535 .

Kousha, K., & Thelwall, M. (2015). An automatic method for extracting citations from Google Books. Journal of the Association for Information Science and Technology, 66 (2), 309–320. https://doi.org/10.1002/asi.23170 .

Kousha, K., & Thelwall, M. (2019). Can Google Scholar and Mendeley help to assess the scholarly impacts of dissertations? Journal of Informetrics, 13 (2), 467–484. https://doi.org/10.1016/j.joi.2019.02.009 .

Larivière, V. (2012). On the shoulders of students? The contribution of PhD students to the advancement of knowledge. Scientometrics, 90 (2), 463–481. https://doi.org/10.1007/s11192-011-0495-6 .

Larivière, V., Zuccala, A., & Archambault, É. (2008). The declining scientific impact of theses: Implications for electronic thesis and dissertation repositories and graduate studies. Scientometrics, 74 (1), 109–121. https://doi.org/10.1007/s11192-008-0106-3 .

Laudel, G., & Gläser, J. (2008). From apprentice to colleague: The metamorphosis of early career researchers. Higher Education, 55 (3), 387–406. https://doi.org/10.1007/s10734-007-9063-7 .

Long, J. S., & McGinnis, R. (1985). The effects of the mentor on the academic career. Scientometrics, 7 (3–6), 255–280. https://doi.org/10.1007/BF02017149 .

Martin, B. R., & Irvine, J. (1983). Assessing basic research: Some partial indicators of scientific progress in radio astronomy. Research Policy, 12 (2), 61–90. https://doi.org/10.1016/0048-7333(83)90005-7 .

Nakagawa, S., Johnson, P. C. D., & Schielzeth, H. (2017). The coefficient of determination \({R}^{2}\) and intra-class correlation coefficient from generalized linear mixed-effects models revisited and expanded. Journal of the Royal Society Interface, 14 (134), 20170213. https://doi.org/10.1098/rsif.2017.0213 .

Nederhof, A., & van Raan, A. (1987). Peer review and bibliometric indicators of scientific performance: A comparison of cum laude doctorates with ordinary doctorates in physics. Scientometrics, 11 (5–6), 333–350. https://doi.org/10.1007/bf02279353 .

Nederhof, A., & van Raan, A. (1989). A validation study of bibliometric indicators: The comparative performance of cum laude doctorates in chemistry. Scientometrics, 17 (5–6), 427–435. https://doi.org/10.1007/BF02017463 .

Niggemann, F. (2020). Interne LOM und ZLV als Instrumente der Universitätsleitung. Qualität in der Wissenschaft, 14 (4), 94–98.

Oestmann, J. W., Meyer, M., & Ziemann, E. (2015). Medizinische Promotionen: Höhere wissenschaftliche Effizienz. Deutsches Ärzteblatt , 112 (42), A–1706/B–1416/C–1388.

Projektgruppe Indikatorenmodell. (2014). Indikatorenmodell für die Berichterstattung zum wissenschaftlichen Nachwuchs . Retrieved from https://www.destatis.de/DE/Themen/Gesellschaft-Umwelt/Bildung-Forschung-Kultur/Hochschulen/Publikationen/Downloads-Hochschulen/indikatorenmodell-endbericht.pdf .

Rogge, J.-C., & Tesch, J. (2016). Wissenschaftspolitik und wissenschaftliche Karriere. In D. Simon, A. Knie, S. Hornbostel, & K. Zimmermann (Eds.), Handbuch Wissenschaftspolitik (2nd ed., pp. 355–374). Berlin: Springer. https://doi.org/10.1007/978-3-658-05455-7_25

Statistisches Bundesamt. (2018). Prüfungen an Hochschulen 2017 . Retrieved from https://www.destatis.de/DE/Themen/Gesellschaft-Umwelt/Bildung-Forschung-Kultur/Hochschulen/Publikationen/Downloads-Hochschulen/pruefungen-hochschulen-2110420177004.pdf .

Vollmar, M. (2019). Neue Promovierendenstatistik: Analyse der ersten Erhebung 2017. WISTA - Wirtschaft und Statistik, 1, 68–80.

Wespel, J., & Jaeger, M. (2015). Leistungsorientierte Zuweisungsverfahren der Länder: Praktische Umsetzung und Entwicklungen. Hochschulmanagement, 10 (3+4), 97–105.

Wissenschaftsrat. (2002). Empfehlungen zur Doktorandenausbildung . Saarbrücken.

Wissenschaftsrat. (2004). Empfehlungen zu forschungs-und lehrförderlichen Strukturen in der Universitätsmedizin . Berlin.

Wissenschaftsrat. (2007). Forschungsleistungen deutscher Universitäten und außeruniversitärer Einrichtungen in der Chemie . Köln.

Wissenschaftsrat. (2008). Forschungsleistungen deutscher Universitäten und außeruniversitärer Einrichtungen in der Soziologie . Köln.

Wissenschaftsrat. (2011). Ergebnisse des Forschungsratings Elektrotechnik und Informationstechnik . Köln.

Wissenschaftsrat. (2012). Ergebnisse des Forschungsratings Anglistik und Amerikanistik . Köln.

Yoels, W. C. (1974). On the fate of the Ph.D. dissertation: A comparative examination of the physical and social sciences. Sociological Focus, 7 (1), 1–13. https://doi.org/10.1080/00380237.1974.10570872 .

Download references

Acknowledgements

Funding was provided by the German Federal Ministry of Education and Research [Grant Numbers 01PQ16004 and 01PQ17001]. We thank Beatrice Schulz for help with data collection.

Open Access funding enabled and organized by Projekt DEAL.

Author information

Authors and affiliations.

Abteilung 2, Forschungssystem und Wissenschaftsdynamik, Deutsches Zentrum für Hochschul- und Wissenschaftsforschung, Schützenstr. 6a, 10117, Berlin, Germany

Paul Donner

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Paul Donner .

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Donner, P. Citation analysis of Ph.D. theses with data from Scopus and Google Books. Scientometrics 126 , 9431–9456 (2021). https://doi.org/10.1007/s11192-021-04173-w

Download citation

Received : 21 January 2021

Accepted : 30 September 2021

Published : 24 October 2021

Issue Date : December 2021

DOI : https://doi.org/10.1007/s11192-021-04173-w

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Early career researchers
  • Ph.D. thesis
  • Doctoral students
  • Doctoral degree holders
  • Citation analysis
  • Research performance
  • Research assessment
  • Find a journal
  • Publish with us
  • Track your research

Turnitin Access To Plagiarism Check For GMS Students

Turnitin is an online plagiarism checking tool that compares your work with existing online publications.

To gain access:

  • Request  access  to the   Plagiarism-Check Blackboard site
  • Access to this site will be continuous throughout your time in GMS

How it works:

  • Upload your papers to Blackboard Learn to check for similarity index
  • Submit  multiple versions of papers, take home assignments, theses or dissertations and rewrite text as needed

Extended Directions:

  • You will be asked to give your name, email BUID and the submission type (i.e. Dissertation, Thesis or Paper) and your GMS program affiliation.
  • After verifying you are not a robot, you be should sent to a new page and get a notification saying “We have received your request to be added to the plagiarism check Blackboard Learn site.” It will take some time for your request to be approved (at least 10 min, possibly 24 hr). The next time you log in to Blackboard, under ‘My Courses’ in the right-hand column you should see the entry:  “GMS Plagiarism Check”.
  • Then click on “>> View/Complete”  (The first time you use Turnitin it will ask you to agree the user agreement). If you have never submitted a document before there should be a submit button.  If you have submitted before you will need to click the resubmit button to submit your new or revised document.  You may get a warning that resubmitting will replace your earlier submission.  It also reminds you that you can only upload 3 documents in 24 hours.
  • You can browse to find the document on your computer then click ‘upload’
  • You will get a message saying it may take up to 2-min to load
  • Once the document is loaded be sure to click on ‘confirm’
  • You should get a congratulatory message.  (You will also get an email confirming the successful submission.)
  • Click on the link to return to submission list
  • You will likely see that your document ‘similarity’ index is ‘processing.’  The algorithm may take a few hours to run.  You will need check back and see when it has finished.
  • Press “view” on the far right to see the submitted manuscript with Turnitin’s Feedback Studio which has some interesting automation features.
  • Alternatively, you can download the results with the arrow icon on the far right of the display
  • NOTE:  Students cannot submit more than 3 documents in a 24hour period.  Please plan accordingly or use an alternatively resource like Turnitin Draft Coach alternative via Google Doc
  • It is recommended that you remove the bibliography/references prior to submission to Turnitin.
  • Because Turnitin cannot scan Images and Figures these can also be removed prior to the check. You must manually check images and figures for plagiarism and potential copyright violations.
  • Published manuscripts should be removed from your thesis or dissertation prior to submission to Turnitin, assuming that they have already been analyzed by Turnitin. If they have not been, they should be included. Unpublished manuscripts should be included in your submission to Turnitin. • How to interpret a “Turnitin Originality Report”

 How to Interpret Your Score Report

  •  Similarity index, Similarity by Source, Internet Sources, Publications, & Student Papers
  • Generally, the similarity index should be less than 20%.
  • A common definition may be acceptable.
  • Did it identify methods or protocols from your lab? These will need to be rewritten.
  • Did it identify matches to publications or text from the internet—again these sentences, paragraphs,  sections will need to be rewritten
  • If you have any questions interpreting the Turnitin report please reach out to your faculty mentors.
  • Once you have made edits it is important to resubmit the document for a final check.
  • The final Turnitin report should be submitted to your mentor and first reader for approval.
  • Having problems with Turnitin:  Reach out to Dr. Theresa Davies for assistance ( [email protected] )

Frequently asked questions

How do i exclude irrelevant similarities from my similarity score.

exclude-source

Exclude similarities : Step 1: Open your Plagiarism Check results. Step 2: Click on the highlighted similarity that you would like to exclude. Step 3: Click on the “Exclude” button on the right.

The similarity is now excluded from your total similarity score.

If you have excluded multiple sources from your total similarity score, then you might see an error message in your Plagiarism Check results. Unfortunately, this is a problem that we cannot fix at the moment.

If you encounter this issue, you can simply ignore all irrelevant similarities and subtract their score from your total similarity score. This allows you to calculate your actual similarity score by yourself.

Frequently asked questions: Scribbr Plagiarism Checker

The free report tells you if your text contains potential plagiarism and other writing issues. The premium report gives you the resources you need to review issues in detail and resolve them.

Scribbr’s free plagiarism checker estimates the risk of plagiarism by calculating the percentage of text in your document that’s similar to other sources.

A moderate or high risk of plagiarism means that the plagiarism software detected several similarities worth reviewing.

Note that similarities are not necessarily plagiarism. You will need to decide on your own whether your text needs revision or citation.

Information can often be found in more than one place. For this reason, other sources citing the same information you used can come up in your Sources Overview.

The important thing is to make sure you’ve cited the source of the material. Try to find the original source, but if you can’t find it, it’s best to cite the source where you found the information.

Similarities in your document are highlighted for quick and easy review. Each color corresponds to a source in your Sources Overview at the right side of your report.

Extensive testing proves that Scribbr’s plagiarism checker is one of the most accurate plagiarism checkers on the market in 2022.

The software detects everything from exact word matches to synonym swapping. It also has access to a full range of source types, including open- and restricted-access journal articles, theses and dissertations, websites, PDFs, and news articles.

No, you will never get a 100% match because the Scribbr Plagiarism Checker does not store documents in a public database.

You can change the settings of your report by clicking on the gear icon which is displayed just above the similarity score.

The settings allow you to exclude search repositories or ignore small matches, quoted material, or references in your bibliography.

Settings plagiarism report

No, the Self-Plagiarism Checker does not store your document in any public database.

In addition, you can delete all your personal information and documents from the Scribbr server as soon as you’ve received your plagiarism report.

Are you not satisfied with the results of the Scribbr Plagiarism Checker , or are you experiencing difficulties with the document? Find out what to do when…

  • you cannot see your plagiarism results: Try opening the results in Google Chrome, or request a PDF copy .
  • you see an error message in the header of your document: Don’t worry – you can ignore the error message .
  • you see “Error! Bookmark not defined” in your table of contents: Don’t worry – this error doesn’t affect your results .
  • you are not satisfied with your similarity score: The similarity score shows you what percentage of your text the software found in sources in the Turnitin database. That means we cannot change your score.However, you can reduce the similarities yourself by following our simple guides to interpreting the report and avoiding plagiarism . Then the similarity score of your submitted paper will be much lower. If you want to see your new score after following these guides, you can purchase a second check .
  • you are not 100% happy with our service: Read about our 100% happiness guarantee and fill in the feedback form. We will contact you within 24 hours.

If you copied a fragment from a source and the Plagiarism Check did not find it, there are four possible explanations.

You paraphrased the fragment

If you paraphrased the original text by using different words and/or changing their order, the fragment will no longer be detected by the plagiarism software.

You can test this by searching the fragment in double quotation marks on Google. If no results are returned, you successfully paraphrased. Note that you still need to cite the source of the original idea.

However, if Google did return a result but our software did not highlight it, this might mean that…

The source is not part of our database

The Turnitin database used by the Scribbr Plagiarism Checker contains 99.3 billion current and historical webpages and 8 million journal articles and books. However, it’s possible that the source you used is an inaccessible publication or a student paper that is not part of our database. Even if you’re certain that the source is publicly accessible, it might not yet have been added to the database. Turnitin is constantly updating its database by searching the internet and adding new publications. You can expect the source to be added in the near future.

To cater to this, we have th e Self-Plagiarism Check er at Scribbr. Just upload any document you used and start the check. You can repeat this as often as you like with all your sources. With your Plagiarism Check order, you get a free pass to use the Self-Plagiarism Checker. Simply upload them to your similarity report and let us do the rest!

The “Exclude Small Matches” number is set too high

By default, the plagiarism report shows similarities of 9 words or more. If the fragment you copied is shorter than that, it will not be shown in the report.

You can manually adjust the minimum length in order to show shorter similarities. Find out how to do that here .

Your document was not readable

If you uploaded a PDF file, it is possible that your document is not machine readable or was converted to an image instead of text. As a result, no similarities will be found.

To test if the text in your PDF is readable, you can copy and paste the document into a text editor (e.g. Microsoft Word, Notepad, TextEdit). If the text editor shows the same text as the PDF, then the text will also be readable by our plagiarism software.

If you know that you used a fragment, but it wasn’t found by our plagiarism software, it’s best to paraphrase or quote it anyway (and be sure to cite the source ). It’s always better to be safe than sorry!

Yes, Scribbr offers a limited free version of its plagiarism checker in partnership with Turnitin. It uses Turnitin’s industry-leading plagiarism detection technology and has access to most content databases.

Run a free plagiarism check

If you’re a university representative, you can contact the sales department of Turnitin .

Scribbr is an authorized Turnitin partner

Scribbr’s Plagiarism Checker is powered by elements of Turnitin’s Similarity Checker , namely the plagiarism detection software and the Internet Archive and Premium Scholarly Publications content databases .

The add-on AI detector is powered by Scribbr’s proprietary software.

Your document will be compared to the world’s largest and fastest-growing content database , containing over:

  • 99.3 billion current and historical webpages.
  • 8 million publications from more than 1,700 publishers such as Springer, IEEE, Elsevier, Wiley-Blackwell, and Taylor & Francis.

Note: Scribbr does not have access to Turnitin’s global database with student papers. Only your university can add and compare submissions to this database.

Scribbr’s plagiarism checker offers complete support for 20 languages, including English, Spanish, German, Arabic, and Dutch.

The add-on AI Detector and AI Proofreader are only available in English.

The complete list of supported languages:

If you purchased a Plagiarism Checker in combination with our Proofreading & Editing service, you can start the Plagiarism Checker anytime at your convenience.

Scribbr recommends using the Plagiarism Checker after you have received your document and gone through all of the edits. Then you can upload the most recent version of your paper and avoid having to manually transfer changes from one document to another.

No, your teacher, professor, or admissions officer will not be able to see your submissions at Scribbr because they’re not added to any shared or public databases.

Your writing stays private. Your submissions to Scribbr are not published in any public database, so no other plagiarism checker (including those used by universities) will see them.

If your university uses Turnitin, the result will be very similar to what you see at Scribbr.

The only possible difference is that your university may compare your submission to a private database containing previously submitted student papers. Scribbr does not have access to these private databases (and neither do other plagiarism checkers).

To cater to this, we have the Self-Plagiarism Checker at Scribbr. Just upload any document you used and start the check. You can repeat this as often as you like with all your sources. With your Plagiarism Check order, you get a free pass to use the Self-Plagiarism Checker. Simply upload them to your similarity report and let us do the rest!

Yes, if you are interested in buying more than 25 plagiarism checks, please fill in this form . We will get back to you within the next two working days!

For questions, please do not hesitate to reach out to [email protected] .

You can download or print your plagiarism report using the “Share” button in the bottom-left corner of your plagiarism report.

download-plagiarism-report

Review every similarity for plagiarism, and decide whether or not you need to revise your text.

  • Review the similarity, and think about whether or not the match makes sense to you.
  • Revise the snippet if necessary. You can do so by paraphrasing or quoting . Always cite your sources.

If you accidentally uploaded the wrong document to the Plagiarism Checker , Scribbr will not refund you for this check or give you a discount for another check. The check starts automatically upon payment, which is why we cannot change your document anymore.

A similarity score of 0% means that our plagiarism software didn’t find any similarities between your document and other documents in our database .

On average, only 1 out of 150 students achieve this. Congratulations!

Your work should not contain any  plagiarism . Even if your score is 1%, you will need to review each similarity and decide whether it’s necessary to revise your work.

But  contrary to popular belief, plagiarism checkers work by detecting similarities, not plagiarism . Not all similarities are plagiarism. Similarities can be:

  • Properly cited quotes
  • In-text citations or your reference list entries
  • Commonly used phrases

What should I do with a found similarity?

Your similarity score shows you what percentage of your text is found within sources in the comparison database.

For example, if your score is 15%, then 15% of the content you wrote is unoriginal, as it matches text in the database.

You will have to review each similarity and decide whether or not you need to revise your work.

What is a good score?   How do I update my work?

If you’re unable to view the Plagiarism Check results in your browser, please try the following two solutions:

1. Are you using Google Chrome?

We’ve found that using Google Chrome resolves most issues related to the results page. First, try using Google Chrome to open your Plagiarism Check results. If you are already using Google Chrome, or if this solution does not work, then proceed to the second solution.

2. Contact support for a PDF copy of the results.

If the first solution doesn’t work, you can contact our support team via chat or email and request a PDF copy of the results via email. This way, you can still review and process the results. We apologize for the inconvenience.

Scribbr checks your document for plagiarism only once. You can use the “exclude text” feature to exclude similarities that you’ve resolved, which automatically updates your similarity score.

If you would like to recheck your entire document, you can purchase a new plagiarism check at the regular price. We don’t currently offer a subscription service.

Sometimes a quote is marked as a similarity by the Scribbr Plagiarism Checker . A quote is not a form of plagiarism , but you do have to ensure you have properly cited the original source.

If you’re certain that you have correctly quoted and cited, you can exclude the similarity from your plagiarism score.

Below we briefly explain the rules for quoting in APA Style .

Quotes under 40 words

When quoting fewer than 40 words, use double quotation marks around the quote, and provide an APA in-text citation that includes the author, the year, and a page number or range.

Quotes over 40 words

If the quote contains 40 words or more, format it as a block quote , which begins on a new line and is indented in its entirety. Include an in-text citation  after the period.

Sometimes it is necessary to quote a source at length:

The Scribbr Plagiarism Checker detect similarities between your paper and a comprehensive database of web and publication content . Because many students write their references in the same way (for instance in APA Style ), a plagiarism checker finds many similarities with these sources.

A reference found by the check is not a form of plagiarism. Therefore, there is no need to take action.

How to exclude your reference list from your results:

You can exclude matches from your reference list in your Plagiarism Check results. However, please be aware that this might not always work, depending on your document.

Step 1 - Deselect Reference list

No worries! This error does not influence the accuracy of your Plagiarism Check results.

If you have inserted a chapter or section title in the header of your document, then you might see an error message in your Plagiarism Check results.

Plagiarism error in header

Instead of the title, you will see an error message like the one above. Unfortunately, this is a problem that we cannot fix.

Our check may flag these error messages as similarities. You can ignore these similarities.

If you have an automatic table of contents and/or list of figures and tables, then you might see an error message instead of the page numbers. The error message might look like this: “Error! Bookmark not defined.”

Unfortunately, this is something we cannot fix. Our check may flag these error messages as similarities. You can ignore these similarities.

Plagiarism check - Error! bookmark not found

The Scribbr Plagiarism Checker is able to work with the following file formats:

The format you use doesn’t influence the final result. If you’re working with a format not listed here, we recommend converting it prior to submitting.

At the moment we do not offer a monthly subscription for the Scribbr Plagiarism Checker. This means you won’t be charged on a recurring basis – you only pay for what you use. We believe this provides you with the flexibility to use our service as frequently or infrequently as you need, without being tied to a contract or recurring fee structure.

You can find an overview of the prices per document here:

Please note that we can’t give refunds if you bought the plagiarism check thinking it was a subscription service as communication around this policy is clear throughout the order process.

By default, the plagiarism report only shows similarities of 9 words or more.

length of similarity

You can change the minimum length of a similarity yourself in order to exclude small matches from the plagiarism report.

What is the best value for the “Exclude Small Matches” setting?

The ideal minimum length of similarities is different per case. In general, we believe that setting the minimum to 8 words results in too many small similarities that are unlikely to be plagiarism. That’s why the default value is set to 9.

However, if you believe that your document contains many small similarities that are not plagiarism ( like these ), you can try changing the setting to 10 or 11 to get a more accurate view of the potential plagiarism in your document.

Proceed with caution , since increasing the number might hide potential plagiarism in your report.

How can I change this “Exclude Small Matches” setting?

1) Click on the gear icon on the top right of the plagiarism report

click-the-gear-icon

2) Fill in your desired minimum similarity length

settings-panel-turnitin

3) Click “Done” to go back to the “Sources overview”

When should I change the “Exclude Small Matches” value?

For most people, there is no need to change this value. However, if you want more control over the similarities shown in your report and you think that most of the small similarities are not relevant, you can increase this number.

Please proceed with caution, since increasing this number might hide potential plagiarism in your report. Decreasing the value is not recommended.

What happens when you decrease the “Exclude Small Matches”?

If you decrease this number, shorter similarities will be included in the report. This means you will see more similarities and a higher similarity percentage.

However, most of these extra similarities are unlikely to be instances of plagiarism. Short similarities are often the result of common phrases that appear in many different sources.

What happens when you increase the “Exclude Small Matches”?

If you increase this number, only longer similarities will be included in the report. This means you will see fewer similarities and a lower similarity percentage.

Since the similarities are longer, the similarities you now see are more likely to be plagiarism.

However, by excluding the smaller similarities, you might miss some instances of potential plagiarism.

Ask our team

Want to contact us directly? No problem.  We  are always here for you.

Support team - Nina

Our team helps students graduate by offering:

  • A world-class citation generator
  • Plagiarism Checker software powered by Turnitin
  • Innovative Citation Checker software
  • Professional proofreading services
  • Over 300 helpful articles about academic writing, citing sources, plagiarism, and more

Scribbr specializes in editing study-related documents . We proofread:

  • PhD dissertations
  • Research proposals
  • Personal statements
  • Admission essays
  • Motivation letters
  • Reflection papers
  • Journal articles
  • Capstone projects

The Scribbr Citation Generator is developed using the open-source Citation Style Language (CSL) project and Frank Bennett’s citeproc-js . It’s the same technology used by dozens of other popular citation tools, including Mendeley and Zotero.

You can find all the citation styles and locales used in the Scribbr Citation Generator in our publicly accessible repository on Github .

IMAGES

  1. Course on Vector Similarity Search and Faiss

    similarity search phd thesis

  2. Thesis VS Dissertation: Differences And Similarities

    similarity search phd thesis

  3. Examples of similarity search. For each query, we show the

    similarity search phd thesis

  4. Effective Tips to Reduce Similarity from the text of a Thesis / Report

    similarity search phd thesis

  5. Billion-scale semantic similarity search with FAISS+SBERT

    similarity search phd thesis

  6. (PDF) Similarity in Ph D Thesis in State Universities is Less Than 10%

    similarity search phd thesis

VIDEO

  1. How To Apply For a UK Student Visa

  2. Similarity : Finding Unknown Lengths

  3. Research, Thesis and project differences and relationships #በአማረኛ

  4. How to Find Reference Thesis

  5. The similarity distance on graphs and graphons

  6. How To Write Your PhD Research Proposal

COMMENTS

  1. PDF High-Dimensional Similarity Search for Large Datasets

    to conduct content-based similarity search, e.g., to search for images with content similar to a query image. How to build data structures to support efficient similarity search on a large scale is an issue of increasing importance. The challenge is that feature-rich data are usually represented as high-dimensional feature vectors, and the

  2. thesis

    A similarity index between 20-40 percent generally means there is a problem unless a large portion of text that should have been skipped was not (e.g., block quotes, reference lists, or appendices of common tables). A similarity index in excess of 40 percent is almost always problematic. You really should not depend on the overall similarity index.

  3. High-dimensional similarity search and sketching : algorithms and hardness

    The certified thesis is available in the Institute Archives and Special Collections. Cataloged from student-submitted PDF version of thesis. Includes bibliographical references (pages 241-255). by Ilya Razenshteyn. ... a new C++ library for high-dimensional similarity search. ' An efficient algorithm for the ANN problem over any distance that ...

  4. PDF Representation Learning for Efficient and Effective Similarity Search

    Similarity Search and Recommendation Advisors: Stephen Alstrup, Christina Lioma, Jakob Grue Simonsen Handed in: April 28, 2021 This thesis has been submitted to the PhD School of The Faculty of Science, University of Copenhagen. Abstract How data is represented and operationalized is critical for building computational

  5. Learning task-specific similarity

    We develop an algorithmic approach to learning similarity from examples of what objects are deemed similar according to the task-specific notion of similarity at hand, as well as optional negative examples. Our learning algorithm constructs, in a greedy fashion, an encoding of the data. This encoding can be seen as an embedding into a space ...

  6. PDF Efficient and Effective Similarity Search in Image Databases

    1.3 Thesis Outline The work of the present thesis is organized in six Chapters and two Appendices as follows: In Chapter 2, we explore the state-of-the-art on similarity search, and interactive simi-larity search in image DBs, pointing out major limits suffered by current image retrieval

  7. PDF Abstract of "Approximation Schemes for Inferring Rankings and

    one has access to several rankings of objects of some sort, such as search engine outputs [Dwork et al., 2001], and desires to aggregate the input rankings into a single output ranking that is similar to all of the input rankings: it should have minimum average distance from the input rankings, for some notion of distance.

  8. Algorithms for Similarity Search and Pseudorandomness

    TY - BOOK. T1 - Algorithms for Similarity Search and Pseudorandomness. AU - Christiani, Tobias Lybecker. PY - 2018. Y1 - 2018. N2 - We study the problem of approximate near neighbor (ANN) searchand show the following results:• An improved framework for solving the ANN problem usinglocality-sensitive hashing, reducing the number of evaluationsof locality-sensitive hash functions and the word ...

  9. Scalable Similarity Search

    The Scalable Similarity Search project, led by professor Rasmus Pagh and hosted at the IT University of Copenhagen, is funded by the European Research Council (ERC).The project runs from May 2014 to April 2019. The team includes 3 PhD students and 3 post-docs. The aim of the project is to improve theory and practice of algorithms for high-dimensional similarity search on big data, and to ...

  10. [1811.03064] Towards a Near Universal Time Series Data Mining Tool

    The last decade has seen a flurry of research on all-pairs-similarity-search (or, self-join) for text, DNA, and a handful of other datatypes, and these systems have been applied to many diverse data mining problems. Surprisingly, however, little progress has been made on addressing this problem for time series subsequences. In this thesis, we have introduced a near universal time series data ...

  11. (PDF) EVALUATING OF EFFICACY SEMANTIC SIMILARITY METHODS ...

    evaluating of efficacy semantic similarity methods for comparison of academic thesis and dissertation texts.pdf Available via license: CC BY 4.0 Content may be subject to copyright.

  12. Plagiarism and what are acceptable similarity scores?

    Plagiarism and what are acceptable similarity scores? The Similarity Report is a flexible document that provides a summary of matching or similar text in submitted work compared against a huge database of Internet sources, journals and previously submitted work, allowing students and instructors to review matches between a submitted work and ...

  13. Learning Task-Specific Similarity

    Chapter 3: Learning embeddings that reflect similarity; This is the main algorithmic "meat" of the thesis: this chapter describes three algorithms for learning an embedding of the data into a weighted Hamming space, with the objective for L1 distance there to reflect the underlying similarity. The algorithms are: Similarity sensitive coding (SSC).

  14. OATD

    You may also want to consult these sites to search for other theses: Google Scholar; NDLTD, the Networked Digital Library of Theses and Dissertations.NDLTD provides information and a search engine for electronic theses and dissertations (ETDs), whether they are open access or not. Proquest Theses and Dissertations (PQDT), a database of dissertations and theses, whether they were published ...

  15. Similarity Index of Doctoral Theses Submitted to Universities in Kerala

    Faculty of Arts (26.6%) has a high degree of similarity followed by Humanities (26%) and lowest in Science (16%).Among the Universities University of Calicut occupied the top position with. 21 % similarity index in its PhD theses followed by M.G University having 20.4% and the University of Kerala has least with 17.9% of similarity index.

  16. Contextual Document Similarity for Content-based Literature Recommender

    context to the similarity. The contextual document similarity is defined as a triple of two documents and the context that specifies to what the similarity of the two documents relates. On a technical level, we will incorporate semantic features from text and links in. hybrid manner to represent documents.

  17. iThenticate

    The similarity check service (iThenticate) can be used by doctoral candidates or their supervisors to assess the work. Find out about the level of similarity with other publications and incorrect referencing before you send (parts of) the thesis to the Assessment Committee, a publisher or send in the thesis for deposit in the UM repository.

  18. Citation analysis of Ph.D. theses with data from Scopus and ...

    This study investigates the potential of citation analysis of Ph.D. theses to obtain valid and useful early career performance indicators at the level of university departments. For German theses from 1996 to 2018 the suitability of citation data from Scopus and Google Books is studied and found to be sufficient to obtain quantitative estimates of early career researchers' performance at ...

  19. UK Doctoral Thesis Metadata from EThOS // British Library

    UK Doctoral Thesis Metadata from EThOS. The datasets in this collection comprise snapshots in time of metadata descriptions of hundreds of thousands of PhD theses awarded by UK Higher Education institutions aggregated by the British Library's EThOS service. The data is estimated to cover around 98% of all PhDs ever awarded by UK Higher ...

  20. A feature-based method for tire pattern similarity detection

    The tire pattern photos are then realized, measured, and detected for similarity following the rules of similarity theory. Experiments were designed to fit the framework of the given detection method. The test results prove that the proposed method can detect tire pattern similarity quickly and conveniently.

  21. Turnitin Access To Plagiarism Check For GMS Students

    Students should search their Blackboard site for "plagiarism" and see if they have access. If not, then follow this link to request access: Plagiarism-Check Blackboard site. You will be asked to give your name, email BUID and the submission type (i.e. Dissertation, Thesis or Paper) and your GMS program affiliation.

  22. PDF Similarity Detection and Plagiarism Verification of M.Phil

    M.Phil. dissertation/Ph.D. thesis. 1. The document must be submitted well in advance, at least 10 days before the deadline for submission of the M.Phil. dissertation/Ph.D. thesis. 2. The document should include abstract, summary, hypos, observations, results, thesi conclusions and recommendations. 3. The similarity check shall exclude the ...

  23. How do I exclude irrelevant similarities from my similarity score?

    You can exclude these irrelevant similarities from your total similarity score. Exclude similarities: Step 1: Open your Plagiarism Check results. Step 2: Click on the highlighted similarity that you would like to exclude. Step 3: Click on the "Exclude" button on the right. The similarity is now excluded from your total similarity score.