Investing additional efforts into sharing data, research outputs such as presentations, or the final published products of the research work, may directly benefit a scientist’s career (especially with the growing attention given to alternative metrics “ altmetrics ”, vide infra ), leading to new collaborations, new funding or even facilitate new discoveries. While there is certainly no shortage of software tools available to share and amplify research efforts, we will discuss a small number of free tools that we use ourselves, of course others are likely to have their own personal favorites. We hope this serves to whet the appetite, encourage further exploration to find out what tools work best for you to meet your objectives, and even form the foundation to discover new or other tools not identified in this piece.

A scientific publication is considered the most basic and historical path to “sharing” the product/s of one’s research. This is also generally considered the final output detailing the purpose for a particular piece of research, provides enough detail to reproduce the work and access to sufficient data (either directly in the manuscript, as supplementary info files, or via links to data stored on other external websites). Nowadays however, even a tweet with links to further scientific information or data can potentially represent a “ nanopublication ”, although the longevity of some of these tools is questionable. This effort does not have to wait until publication as you could tweet ideas at the very earliest stages of the research depending on your level of openness, especially in regards to open data, as well as concern for being scooped by people following you. In the process of our own experiments to determine the benefits and paths towards online sharing of our research, we have identified at least four steps that the reader could put into action. These four steps are to explain, share, and gather evidence of increased awareness of the work, as well as gather feedback from the community. These efforts may only take a few minutes of a scientist’s time, which is a fractional investment compared with the hundreds to thousands of hours spent on the scientific research from conception to publication. These steps, in turn, may increase impact of the research efforts.

Based on our collective experiences over the past five years, we believe there is a significant return on investment, and it is far higher than simply relying on publishers to invest informed and targeted efforts in sharing a scientific publication. This in fact rarely happens unless the publisher, or you, invest in a press release, but the vast majority of publications are certainly never picked up by the general press, primarily due to the sheer volume of scientific papers being published. For a publisher producing many thousands, if not tens of thousands of publications in a year, investing efforts to share information about one publication above and beyond one tweet, or a rudimentary blog post on their website, is highly unlikely. If publishers were to promote each article, book or report they publish, this would flood their communication channels, thus creating too much noise. Their own marketing efforts would be less effective through mass communication and may also reduce reach, especially to the overwhelmed specific target audiences that would benefit from the author's research. Ultimately, the best person to raise awareness is the author(s) themselves who can leverage their own networks, whether through electronic tools or personal interactions. These may be as effective as a press release when it comes to spreading the word on what is likely a hot topic relevant to few scientists in a specialized field.

The vast majority of researchers depend on research funding in order to progress their science. However, acquiring research funding continues to be a challenge, with only 20.7% (14,457/69,973) of NIH Research project grants funded in 2015. This increased competition drives requirements to assess the quality and volume of research output. These exercises are becoming more commonplace, for example the Research Excellence Framework (REF2020) is the current system for assessing the quality of research in UK higher education institutions and, similarly, the Excellence in Research in Australia project (ERA). While there are various subjective judgements regarding the “impact” of a publication (including impact factors , CiteScore 7 , h-index and citations) for a research effort and ultimately for the performance and impact of a scientist, it should be obvious that in our present time the sharing, networking and outreach of research work could bring benefits. What those benefits actually are, and how to deliver them reproducibly, is one of the mysteries of this new scientific networking ecosystem. How does one use technology to maximally increase the visibility of your research outputs and potentially lead to breakthroughs that benefit humanity? This could be seen as the ‘new alchemy’ of science. Suggested benefits might be broader reach for a publication and Altmetric scores ( vide infra ) that could be used by granting agencies or others for assessing the scientist. By engaging or participating in online media for example, scientists can be empowered by using and adopting the tools available to them to share and improve access to their research, facilitating networking and potentially resulting in greater impacts on their career. However, these benefits come at a cost as they require time and effort that could distract from other efforts, such as research, writing grants, reviewing, mentoring and so on, and they are not yet in the standard workflow of the scientist.

This has many potential implications, one being a way to extend the life of a scientific publication and its enmeshing in a network of other electronic outputs. Another is the considerable effort that would be required to fulfill even a sampling of these approaches. Examples of the increasing importance of these non-traditional sources of scientific information is the increasing prevalence of links to Wikipedia, blog posts and code-sharing in reference lists associated with references in scientific publications ( ~35,000 citations to Wikipedia, ~11,000 for YouTube, ~10,500 for Facebook, and ~7000 for Twitter ). This will increase in the future.

With such a large quantity of scientific research finding its way into published works, it is hard for scientists of any description, whether chemists, biologists or physicists to make sure their work rises above the ‘noise’. This is of course important if it is to be seen by peers and in turn used by them and cited, to ultimately be captured and further used to infer the importance for any given publication through various citation metrics. Historically, publishing was seen as the end result of the scientific endeavor, which one might consider as a simplistic linear process from hypothesis – experiment – publication ( Figure 1 ). However, the traditional approach of a particular piece of research dissemination reaching its conclusion once it has been published was a limitation of printed journals and should not pose such limitations in an electronic online world. Now, with the advent of internet-based technologies, the last decade has seen an explosive growth in additional forms of media for the dissemination of publications and associated research data, analyses and results, including wikis, blogs and code-sharing platforms (e.g. GitHub ). The process of research publication is now non-linear, with a potentially infinite variety of dissemination steps that could be taken between hypothesis to publication and beyond ( Figure 1 ). This could include early release of data and scientific manuscripts as preprints. Sharing details regarding a research study and associated data offers a number of potential positive effects that can contribute to the quality of science. The historical approach of peer review was to hopefully both improve and ensure the quality of the science and the published output. Sharing research data and preprints allows for early feedback on both the results and preliminary findings from peer groups and reviewers.

As the world has changed in the past three decades, scientific publishing, seen increasingly as the currency of science for at least 50 years, has also seen a dramatic shift. Since publishing of the first modern journal in 1665 (Le Journal des Sçavans), a staggering number of scholarly articles have been published, reaching the milestone of 50 million by 2009 2 . Recently Ware and Mabe 3 projected between 1.8–1.9 million scientific articles are now being published every year, while contributing one aspect of Big Data for science 4 . In recent years, there has also been an explosion in the number of new Open Access publishers, a number of which seem to be focused on the profit potential of this new marketplace as predatory open access publishers . This exploitative open-access publishing business model charges publication fees without providing the editorial and publishing services associated with legitimate journals. This could allow scientists to game the system by publishing even poor quality science in predatory journals, effectively padding their resume. This was tested in an elaborate sting when a spoof research article was accepted by dozens of open access publishers 5 . It even extended to allowing fake editors roles on their editorial boards 6 . Such dubious practices in the publishing world, especially in terms of gaming a publication record can unfortunately extend to the use of online networking and sharing tools to boost online profiles and game altmetric scores ( vide infra ).

With new tools being developed, it is difficult to keep track, determine what works, and optimize their use, especially in the context of science. For example, Kramer and Bosman (2017) manage a growing list containing over 400 tools and innovations in scholarly communications, and some of these can be used for online sharing. This list reinforces the breadth of software tools available, and demonstrates how this can be confusing for any scientist to know where to start.

In the past 40 years, society has undergone considerable changes driven by the development of affordable personal computers, the internet and most recently, mobile devices, which allow widespread connection to the internet. In turn, these technologies have shaped how we interact with each other and form online networks. Since 2000 there has been an almost 800% increase in the number of people using the internet, with over 3.5 billion people online. Online sharing now ranges from posting opinions, 140 character nuggets on Twitter, updates and discussions on LinkedIn, photos or videos via free platforms such as YouTube or Vimeo, presentation slides on SlideShare and so on. The specific online media sites preferred in each country across the world differs. Despite the prevalence of social media tools, the vast majority of scientists do not use these tools to help share, evidence and amplify their scientific research 1 . We believe there are a number of reasons that most scientists do not use these tools. It could be because few people would even think of using online media tools for their scientific research, or because they do not understand the potential value. The lack of credit for sharing pre-published data, code or other forms of research outputs, especially in terms of citations that can contribute to career progression, may also be an issue. Maybe scientists do not see this activity as a valuable use of their time or they require initial guidance to help navigate use of these tools. Other issues include fear of being scooped, and the idea that if your work is solid, it should be able to stand on its own merits without needing to be “amplified.”

To extract the most value from these efforts it is certainly important to identify your primary goals and objectives for using these networking, sharing and amplification tools. For example, you might be interested in: 1) saving time by using integrated tools; 2) sharing your work using a small number of online services for dissemination and amplification; 3) tracking and providing evidence of your “impact” on research to potentially help with research funding applications; and/or 4) furthering your outreach to those outside your own network and perhaps engaging with the general public. Once you have defined your focus, you will be better positioned in terms of deciding what software services to use and whether to utilize a mixture or concentrate your efforts on one or two only. Your institution, colleagues, collaborators or even publishers may also have some recommendations for you or may already be partnered with specific services of which you can take advantage.

Categories of tools: Networking, sharing, tracking and research amplification

We separated these software tools into four specific categories: networking, sharing, tracking and amplification. We acknowledge that the lines between many of these are actually rather fuzzy and that networking sites, such as Facebook and Twitter, are primarily used for sharing. Many of these tools actually serve more than one particular function, and our article is acknowledged to be subjective and based on our own experience and usage. There are obviously more generic tools that have long existed, for example email, and newer tools like Slack or Yammer, which are generally used for private communication and sharing. While these newer tools may be important for science and collaboration within small groups, they are useful in extending your online network more broadly. Clearly, the changes we have seen with technology may not predict what we will see in future in terms of communication style and utility as technologies themselves can redefine communication styles, as is the case with the 140 character tweet. Throughout this article we will point to examples of our own profiles on these various online tools as examples.

Networking LinkedIn. While LinkedIn has a primary role to form connections and expose your career to potential employers and partners, it is likely the de facto networking tool for many professional scientists. LinkedIn is also the number one networking site for business and crosses all domains. The networking of scientists, business and investors offers potential opportunities for innovation and commercialization. Other platforms (e.g., ResearchGate) are more geared towards research. LinkedIn offers both free and paid levels of service and, to date, we have only used the free services. We think that anyone creating a LinkedIn profile should invest a minimal effort in terms of adding a “professional” photograph, a career history back to their research training (i.e. college or university), a minimum of two to three research publications, and links to public presentations on other sites (vide infra). We also suggest that career interests be listed and some efforts invested in building out a network of work colleagues, ex-supervisors, co-authors, etc. We believe that LinkedIn be considered as an opportunity for career-networking, not family event and activity sharing, and to maintain the highest level of professionalism on this site. The ‘networking effect’ will likely result fairly quickly in new contacts reaching out to join your network, and we suggest being somewhat selective in accepting these offers. Some proposed filters for accepting new contacts include: 1) whether you have met them face-to-face; 2) the size of overlapping networks (for example when you have ~10 contacts in common); and 3) when you have already had either phone or email contact with the person. As well as recommendations from those you have worked directly with, the endorsement facility of LinkedIn allows members of your network to identify and endorse you for specific skills and this, in particular, can be valuable in terms of having a publicly acknowledged list of capabilities visible to the LinkedIn user base. An example of such an endorsement list is shown in Figure 2, which shows a partial list of the featured skills and endorsements for one of the authors (SE). This list can be pruned according to what skills you would want to be identified as having. It is also possible to directly list projects that you have been (or are) involved with, as well as associating the various members of the project team. This further defines the networking aspect of the site. For those who use the site, they will have noticed business managers sharing their greatest coaching ideas or meaningful quotes. A scientist can use the same facility to post an update regarding present activities or simply something of scientific interest to share with followers. In this regard, the site can be considered as a more expansive version of Twitter. Some of these types of updates can be incredibly useful as tools to highlight a recent achievement for your group or collaborators. From our experience, a tendency towards keeping them brief and adding an image certainly helps in obtaining more views and ‘likes’. We have used this facility to provide updates on new papers or grants received as well as updates on specific projects. We have found that positive or engaging news can quickly gather momentum (e.g. announcing new grants, new hires, new jobs) and can drive contacts that lead to offline follow up. As an example, AJW posts updates on LinkedIn regarding his most recent project, the CompTox Chemistry Dashboard8. The associated Google Analytics for the site shows that these postings are very effective in driving traffic to the site, commonly a few dozen visits within a day of posting an update about the site. This can be informative about your network and perhaps beyond, in that good news travels fast. One or more scientific publications can be associated with the profile to illustrate latest research efforts and we suggest association of your highest profile publications with the site. Ideally LinkedIn would make use of Digital Object identifiers (DOIs) and would allow a resolver service, such as CrossRef, to display the publication details rather than forcing manual entry as it does at present. DOIs are one of the primary ways that scientific online tools integrate their data streams (i.e. ORCID, Altmetric, Publons). Besides using LinkedIn to share links to your latest publications, one can also insert PowerPoint presentations, PDF files and other document forms using the embed functionality available via SlideShare (described in more detail later). The regularity of updates regarding new publications and presentations can be representative of your productivity, and requires active attention to your LinkedIn profile. The sharing of articles, presentations and science news that interests you can also help drive attention to other people’s work and elevate interest in it. For sharing via the site, we suggest using URL shortener services like TinyURL, Goo.gl URLs or Bitly to track the number of times people click on your link (as clickthroughs). However, it should be remembered that some services are currently banned in countries like China (e.g. Google), so you may want to ensure you are using “global friendly” services for this purpose. In our opinion LinkedIn is a pivotal tool and the purchase of LinkedIn by Microsoft9 would indicate that it will be more closely integrated with their products, maybe even extending to include an integration to their Microsoft Academic site that operates in the same space as Google Scholar Citations. ResearchGate and Academia. There are several tools for networking your publications. Two of the most popular are ResearchGate and Academia. While there are differences in functionality, both sites provide similar abilities in terms of sharing preprint manuscripts, presentations, posters and other forms of general communication of your science. For example, one author’s (AJW) Academia site lists ~250 publications, presentations, book chapters, magazine articles and other research outputs and has had ~30,000 views. The detailed analytics page includes geographical details, views and downloads and Academia’s measures of impact). A recent study in PLOS ONE10 states that papers uploaded to Academia receive a 69% boost in citations over five years. ResearchGate appears to be more expansive in terms of what can be hosted on the site and can include, for example, datasets, project updates, patents, working papers, code, and negative results. Users are encouraged to fill out their complete profile, and list awards and their previous work experience at a minimum. It also provides a forum for technical questions and answers, and most recently, allows users to group publications and data into ‘projects’. SE’s profile includes 418 contributions as of 19th March 2017 and provides a reach of ca. 15,000 researchers, who can learn about his publications, presentations and postings to the site. It should be noted he also barely uses this software, as he prefers to put his efforts elsewhere. Academia and ResearchGate, when used to develop a network of followers, both offer an excellent communication path for sharing activities and research outputs with the community using the sites. These sites require some considerable effort to respond to requests and engage with others if you are to fully maximize their value. It is important to note that any uploads of published work to these services requires permission from the publisher. A strong movement from publishers, which is beginning to take momentum, puts pressure on these services to ensure they educate their users on appropriate copyright. You may find the publisher will contact you if they feel that your published work is listed on ResearchGate and conflicts against your agreed copyright. To ensure that ResearchGate is used without risking a breach of copyright, it is necessary to read the Copyright Transfer Agreement or transfer license associated with the publisher to see how the paper can be shared. Open access policies can be checked on the journal’s website or by using SHERPA RoMEO, a site that presents policies based on a journal title or ISSN search.

Sharing Currently, the most widely adopted social sharing platform is Facebook, approaching two billion users as of May 2017. While we acknowledge the penetration, versatility and general global acceptance of Facebook as a platform, we collectively use it for sharing with friends and family mainly, only rarely sharing links to some of our scientific works. We prefer to separate career activities from personal pursuits, but acknowledge that this is also a personal choice. For younger generations this may be inverted, their followers on Facebook may be colleagues rather than family and for them it would be seen as appropriate to share on Facebook, on Instagram or on group chat site, such as WhatsApp. Instead, for sharing we use alternative online tools as outlined below. Blogs. All three of us have managed blogs for a number of years and two invested considerable amounts of time in sharing opinions, data and ideas (SE and AJW). We have developed collaborations of value, asked for opinions and guidance and used them to share information regarding our (or others) latest research efforts, industry research and news/updates. Overall though, it seems that there is less interest in blogging, probably due to the amount of time needed to develop content and we ourselves more commonly share information now in smaller soundbites (via Twitter and LinkedIn). We also try to reach other networks by being guest bloggers on the blogs of others or news services and using these other sites to share data. There are so many other applications now available for sharing relative to just a few years ago that we will therefore focus our discussion on those sites that we use on a more regular basis. “Small nugget sharing”: Twitter/Google Plus. In many ways both Twitter and Google Plus are for sharing bite-sized communications, presently limited to only 140 characters for Twitter, but possibly to be extended. In terms of sharing, it is an ideal application for pointing your network to papers, publications, data sets, and blog posts via embedded short links. Networking retweets can drive attention to these various types of resources. It is simple to use and push out a URL to something you wish to share can drive attention easily, though only for a fairly short period of time. Once an item of interest has been prepared on some other form of social media, or a website link to a publication requires sharing, Twitter and Google Plus then become ideal ways to point to these items. Examples of nugget sharing include letting people know in advance that you will be at a conference (and potentially set up a “tweetup” with new connections), or live tweeting at a conference11. One of the primary reasons that we use these platforms is for amplification of our publications (vide infra). Both AJW and SE have initiated collaborations and specific data sharing activities via Twitter. For example, SE attended a meeting and shared the need for an application to build on a Green Chemistry dataset that was available only as a PDF file. What was initiated as a public Twitter exchange resulted in a mobile application being available within a few days12, an example of a specific collaboration initiated via 140 character exchanges. Another example is a Twitter exchange that gained support for a new chemical identifier associated with the EPA CompTox Chemistry Dashboard added to Wikidata. Also, an open research project on Ebola (SE) was initiated, which in the space of two years facilitated the identification of new antivirals, publications13–17 and an eventual NIH grant. Others have described how Twitter can be used by scientists to extend network networks and ultimately find jobs18 and, in our domain of chemistry, for sharing “Real-time Chemistry”. Taking advantage of the communities of followers on Twitter and Google Plus is only possible, of course, once you have established a community. Building a community requires engaging in the platform by following other users who post interesting content, by engaging with the content, sharing other people’s content and posting your own. Developing your own following is an incremental process, which may take years and there are more expansive guidelines available on how to do this. Media sharing: Presentations, preprints and videos. Most scientists present their work at conferences, either as talks or posters. Without using online sharing tools, the only people who would see your presentations, which commonly take hours of time to put together, would be limited to the people at the meeting. However, online tools for sharing and distributing these same presentations can result in a much broader reach, and importantly, keep the work alive for a period much longer than the limited presentation time, and associated audience, at the conference. One commonly used presentation sharing platform is SlideShare (acquired by LinkedIn in 2012). An advantage of these websites is that they both belong to the same company (Microsoft). This allows a relatively simple process for associating a presentation with your profile with one click (“Add to Profile”) to make the presentation visible (Figure 3). SlideShare is not limited to simply sharing PowerPoint presentations: the user can include article preprints, infographics and other documents, as well as the ability to have integrated videos embedded via YouTube. One approach (adopted by AJW) to derive most value from the network effect of multiple connected platforms, reach the broadest audience, and share the work in various forms is as follows: 1) a PowerPoint presentation delivered at a meeting is shared on SlideShare (and also figshare, ResearchGate and Academia); 2) a narrated version with voiceover to capture the presentation is made and published to YouTube; 3) the YouTube embed function is used to insert the video to the second slide on SlideShare; and, 4) a viewer will then have the choice to view the slide deck, download it for local storage, and if they want they can hear the author also present the slide deck with the voiceover. While we acknowledge that the most popular global video sharing platform is YouTube, there are geographical issues, not only based on language, in terms of all countries accepting the sharing platform. Streaming content in China via YouTube is an issue and other platforms, such as Vimeo or Weibo, may be an option. We think that scientists should not mix their scientific movies (for example, narrated presentations, lab activities, etc.) with family movies! Data sharing. There are myriad platforms available for data sharing, and it is difficult to be exhaustive in this short article as, depending on the particular domain of science, there will be biases. Climate science, versus chemistry, physics or medical sciences have their own favorite platforms. We have experience of the Dryad Digital Repository, figshare and Mendeley data for sharing most data types. Other sites can be used for specific types of data sharing. For example, PubChem for sharing BioAssay data. For the purposes of this article we focus only on the figshare site for data sharing, as we have the greatest experience in terms of using this platform, as well as the fact that it is now integrated with many publishers that we have published our articles with (i.e. PLOS, the American Chemical Society, Springer and Wiley). Importantly, figshare offers the advantage of creating DOIs that give unique persistent identifiers that can then be resolved across platforms. Datasets published to figshare can be embargoed, cited in a manuscript, and made open at the time of publication. This provides important benefits as now specific datasets are fully citable (via DOI), the number of views and downloads are directly tracked, the altmetrics can be measured and, overall, there is significantly more insight into how data is used and accessed over simply putting a file as supplementary info with a publication. An example of a shared dataset, including views, downloads and an associated “Altmetric donut” is shown in Figure 4. The donut represents the Altmetric Attention Score and is designed to identify how much and what type of attention a research output has received. The colors of the donut represent the different sources of attention and the details regarding how the score is calculated are available online. With the data (or presentations or documents) on figshare, we can then share details via Twitter and use the associated DOI to cite our datasets again on ResearchGate. figshare also allows us to share figures before they are used in manuscripts, define them with a CC-BY Creative Commons license, and then use them in our publications. In this way, we are not transferring copyright of our figures to the journals and other people who wish to use our figures do not have to request permission as they are already in the public domain. figshare is also a platform where we share our presentations, posters and preprints (other preprint servers do exist, such as arXiv, bioRxiv, and the impending ChemRXiv. It is also where we tend to host our data that we point to from our publications, as well as using subsets of the overall data in the supplementary information with publications. Tagging of any form of information that we share via the site makes it more discoverable, and a search across a specific application website can therefore surface this to interested parties. More recently figshare has added ‘collections’, and we have made use of this around our work on the Zika virus. As multiple academic publishers have now accepted figshare as a data repository of choice, its reach appears to be growing and it is likely to have increasing importance as a data repository for science. Code sharing. It is advantageous to those that produce computer code, as a part of their scientific output, to be able to share it and allow others to consume it. Major code repositories, such as GitHub and SourceForge, offer many advantages for collaborative software development and versioning, so have great utility in their own right. They are integrated into online sharing platforms, thereby ensuring that code updates are transmitted across the community, keeping the audiences, who may require code access, informed of new depositions that they can consume. Most scientists may never use these repositories, but they are of central importance in the science ecosystem and other data sharing tools can learn from them. There are other reviews regarding the adoption of these platforms and readers are pointed to these for more detail19. We, like many others, believe that code used in projects that is then used to deliver software underpinning, for example, data processing, analysis and reporting, should be citable and ultimately be a part of the altmetrics feeding a scientist’s acknowledgement for their contributions to science.

Impact tracking As scientists, one of our interests is to track our publication records, have access to our citation statistics, and, potentially measure the impact of our work. Impact can be estimated by a number of statistics, such as the h-index. There are even programs available so you can generate your own statistics. In recent years, catalyzed specifically by the work of Priem et al and the release of the “altmetrics manifesto”, altmetric statistics have started to gain general acceptance within the community as measures of interest in a scientist’s work. This acceptance is likely to increase as their algorithm becomes more mature and produces increasingly relevant results. Altmetric statistics not only take account of standard publication metrics, such as citations, article views and downloads, may also track views and downloads for presentations and videos, and include measures of attention for discussion (via blogs) regarding publications online on platforms such as Twitter and Facebook. Altmetric statistics may also attempt to measure the impact of the reuse of data sets and code. This, to many, may seem incredibly complex and go far beyond just tracking “the paper” itself, but this is the world that is evolving. A number of tools attempting to integrate and track these altmetrics impact statistics have been established. These include ImpactStory, Altmetric and PlumX. This section discusses some of the sites we use for managing our publication records and tracking our classical citation statistics, as well as those that we use for measuring our impact in the realm of altmetrics. For the purposes of “publication and citation tracking” we use: ORCID, Google Scholar, and Microsoft Academic Search. While other platforms, such as ResearchGate, do an excellent job of informing us via emails when there are new publications to associate with our profile, once we have confirmed the associated publication as appropriate, we use it more as a networking site and catch-all for the bulk of our research outputs (see earlier). These three sites are more focused on simply tracking publications and, in the case of ORCID, it expands to include presentations listed on figshare (via the DOI integration). Each of these websites is free to use for an individual scientist, while ORCID also offers access to an institutional package (vide infra) that allows organizations to mesh together contributions for their staff into an institutional representation of activity. ORCID. An Open Researcher and Contributor Identifier (ORCID iD) is a unique numeric identifier for a researcher that is free to claim and can be obtained simply by registering at the website. Almost four million identifiers have been claimed (or issued) at the time of writing. The ORCID iD is a derivative of previous efforts by Clarivate Analytics (formally Intellectual Property and Science business of Thomson Reuters) to produce the ResearcherID that has a distinct benefit of disambiguating authors and their association with publications, which is an issue for other sites (see later for a distinct example of this problem with Google Scholar). Once a scientist has claimed their unique identifier, they are responsible for defining the content associated with it, including whether they wish the content to be public or private. They can add a short biography and associate a number of their websites and public profiles with the ORCID site. Increasingly, these identifiers are expected or accepted by publishers at the time of manuscript submission, and funding agencies are also starting to use them. The website allows for an online resume to be assembled from publications by searching based on your name and editing the list as necessary. The data collected are then available via an application programming interface and, for example, can be used by publishers to use on their own platforms for enhancing the linkages between an authors’ publications. As a starting point, it is possible to upload a list of publications in a standard format, such as Bibtex or EndNote. It is also possible for a scientist with a ResearcherID to connect and migrate the existing content to ORCID and expand from that point. The ORCID application programming interface and authorization module allows connectivity between web-applications. Since the ORCID iD itself has value independent of the capabilities of the website as a representation of a scientist’s publication record and resume, obtaining an ORCID iD is, in our estimation, one of the primary entry points into the scientific networking regime today and we encourage registration. Inclusion of the ORCID into PowerPoint templates for presentations shared online, and on other scientific networking and data sharing sites ensures that a simple web search in the future will aggregate the majority of your public works labeled with the identifier. We also add the ORCID iD either on the first or last slide of our slide decks with the intention that it is captured by the search engines, allowing a simple search to provide us a list of ORCID indexed works. Google Scholar and Microsoft Academic Search. Google Scholar (GS) is a free website for assembling what is effectively a list of your publications and their citations. GS also provides metrics, such as the h-index, that we have found to be generally much higher than in the Web of Science (WoS), which could be because it includes self-citations, whereas WoS does not. There have been comments that GS is a better predictor of actual citation numbers than WoS. GS is useful for searching for publications and perhaps picking up citations that the commercial tools are missing. We have seen a worrying recent trend in terms of auto-associating publications, and recently AJW identified that 70 publications had been associated with his GS profile and had to be manually deleted. Profile maintenance is therefore necessary by the user, and there has to be careful curation and pruning on an ongoing basis. Microsoft Academic is much like GS, but we have found it to be less useful because it was unable to capture citations to our publications. “Alternative metrics”: Plum Analytics, Altmetric, ImpactStory. These tools aggregate the citations from blogs, tweets, Facebook, etc., and use their own algorithms to derive a score for each paper. Twitter is commonly a major contributor in terms of social media counts, and a weighted approach in terms of the importance of a social media event can be taken into account. For example, Altmetric give a news item a higher weighting than a tweet. Interestingly, ResearchGate also derives a score for each author, though it remains a little unclear how the score is calculated.

Amplification Kudos. The emerging area of author support tools has very limited research findings available (e.g. data and code). However, Kudos has been recently highlighted as useful in this regard20. While there are plenty of websites for an author to post their papers and preprints, enriching these research outputs to add more information about what commentaries have been made about the work, linking to additional presentations, datasets, etc., is less supported in general. Kudos tracks citations (supported only by WoS statistics at present), altmetrics on publications (supported by Altmetric currently), and other statistics, like usage, where available (e.g., downloads and clickthroughs). It also provides a dashboard of an author’s papers using a CrossRef DOI as the basis of the data feed and can be linked to your ORCID account (Figure 5). For a new user there is a downside if they have a large publication record, as it will take a very long time to enrich every publication with links to other content, but a user can of course choose to ignore their historical record of publications and focus only on new publications moving forward. Our adoption of the platform has allowed us to enrich publications with information (with examples described below) and share the associated Kudos page, drive traffic to the paper and track this activity. In our experience, Kudos results can be improved when multiple authors contribute and work together to improve awareness of a publication because the publication is disseminated through multiple-author networks and online networking efforts. The enrichment capability offered by the Kudos platform delivers a valuable capability to the authors of the publication: the ability to keep the publication up to date by linking to related information, for example, blog posts regarding the work, media coverage, presentation slide decks from conferences, later publications by the authors themselves or derivative works by other scientists. One of the authors (AJW) coined the term “forward citation” to refer to this capability, as citations are retrospective in nature and point only to earlier work. Enrichment of an article can continue to keep the research reported in a publication updated with follow-on information as later works of various types are associated. An example of this from the author’s list of publications is associated with the synthesis of the chemical known as “Olympicene”, a small molecule synthesized during the time of the 2012 Olympic Games as a form of molecular commemoration. The claimed publication has been enriched with a YouTube movie, multiple blog posts, various types of media coverage and even detailed discussions regarding part of the chemical synthesis. In particular, multiple scientific papers by other authors referring specifically to the trivial chemical name of Olympicene have been associated with the original paper thereby connecting the work directly. While search engines and referential systems from publishers attempt to do this in an automated manner, in this case the claiming author(s) have an opportunity to directly make the association and comment as appropriate, regarding the value and nature of the related information. In June 2016, Kudos and the Altmetrics Research Team at the Centre for HEalthy and Sustainable CitieS (CHESS), and the Wee Kim Wee School of Communication and Information at Nanyang Technological University (NTU, Singapore) analyzed user data and found that 51% of registered Kudos users were STEM researchers sharing their work, 29% were social sciences researchers and 8% were humanities researchers. This demonstrates the engagement of STEM researchers, and how they are leading the way in using innovative tools to disseminate their research. However, with seven to nine million active researchers a year3, Kudos only has around 1% penetration into the research community, so there is huge growth potential here for these types of services to help more of the scientific community.