Computing Profession – Communications of the ACM https://cacm.acm.org Mon, 10 Mar 2025 20:32:41 +0000 en-US hourly 1 https://wordpress.org/?v=6.7.2 https://cacm.acm.org/wp-content/uploads/2023/11/cropped-cropped-cacm_favicon-1.png?w=32 Computing Profession – Communications of the ACM https://cacm.acm.org 32 32 212686646 Beyond Compliance: Security Documentation as a Strategic Asset https://cacm.acm.org/blogcacm/beyond-compliance-security-documentation-as-a-strategic-asset/ https://cacm.acm.org/blogcacm/beyond-compliance-security-documentation-as-a-strategic-asset/#respond Mon, 10 Mar 2025 15:48:54 +0000 https://cacm.acm.org/?p=765812 It’s time to stop viewing security documentation as a necessary evil and start leveraging it as a strategic asset.

]]>
For many organizations, security documentation is the chore that ticks a compliance box, a necessary evil that soaks up time and resources without offering apparent value. 

But what if it could be more than a compliance exercise? What if it could serve as a powerful strategic asset, driving operational excellence, fostering organizational resilience, and even becoming an essential aspect of cybersecurity culture?

Hence, treating documentation as a static repository of policies is no longer enough. Forward-thinking organizations are reimagining security documentation as a living framework—an essential tool that shapes decision-making, aligns teams, and fortifies both technical and business strategies.

Avoid a Compliance-Only Mindset

Most organizations approach security documentation with a narrow goal: passing audits and avoiding penalties. This compliance-first mindset tends to generate documentation that is reactive, fragmented, and often outdated by the time it’s reviewed. Policies become shelfware, existing in neatly organized digital folders but disconnected from daily operations.

The result? A false sense of security. Documentation, no matter how robust it really is, then exists only to meet regulatory requirements. Furthermore, it doesn’t necessarily reflect the dynamic nature of modern cyber threats. 

It doesn’t engage employees, nor does it provide actionable guidance when an actual security incident occurs. So, how do we bring it back to life? 

Reframe Documentation as a Living Asset

Treating documentation as a strategic asset means integrating it into the operational fabric of your organization. Rather than static PDFs gathering digital dust, security policies, procedures, and guidelines should evolve continuously alongside technological shifts, business objectives, and threat landscapes.

This dynamic approach transforms documentation from a reactive tool into a proactive enabler, allowing it to become a mechanism for:

The Anatomy of Strategic Security Documentation

What differentiates strategic security documentation from compliance-driven paperwork? A strategic approach has an entirely different motive—prevention and facilitation. As such, you’ll notice the right security docs being characterized by: 

Clarity and Accessibility

Security policies should be written in plain language, avoiding unnecessary jargon that might alienate non-technical staff. Furthermore, documentation must be stored in easily accessible formats, using digital platforms that allow employees to quickly locate the information they need during critical moments.

Relevance

Documentation must remain in sync with technological innovations, regulatory shifts, and internal business changes. Thus, regularly scheduled reviews should be conducted to ensure that the content reflects current realities, not outdated assumptions. The reviewer needs to be a third party, lest you want to go back to a compliance-first approach. 

Effective Localization

This is especially important for international companies. While you can use a translation API to crunch the text, make sure any documentation is reviewed by humans before distribution. Don’t rely on everyone’s English proficiency and the ability to understand complex cybersec jargon. Be direct and speak to each region directly. 

Integration with Business Goals

Security shouldn’t be seen as a roadblock to innovation. Instead, well-crafted documentation helps organizations balance risk with progress by aligning security policies with broader business objectives. When security goals align with business strategies, innovation can thrive without compromising data integrity.

Embedding Documentation into Organizational Culture

Making security documentation a strategic asset demands a cultural shift within the organization. Simply rewriting policies isn’t enough; businesses need to weave a documentation-first mindset into everyday operations. This begins with leadership. Senior management must champion dynamic documentation as a strategic priority, setting the tone for the rest of the organization.

Engaging employees is equally critical. Regular training sessions and clear communication help staff understand their roles and responsibilities within the security framework. Take healthcare and its essential role in our society as an example. 

Given the fact that healthcare facilities experience over 700 cyber attacks a year in the U.S. alone, this is crucial for them and any other vulnerable industries. But it goes further than just having HIPAA-compliant hosting and having nurses not click on suspicious links.

On the backend, technology also plays a pivotal role. Implementing tools that allow real-time updates, maintain version control, and offer easy access to documents ensures that the documentation remains up-to-date and usable. 

So, if you include these practices in your organization’s culture, you can transform documentation from a static requirement into an evolving, strategic asset that supports growth and resilience.

Metrics for Measuring Documentation Effectiveness

How do you know if your documentation strategy is working? The exact answer varies on a case-by-case basis, but generally hinges on the efficacy of your: 

  • Incident response time: A key performance indicator (KPI) for the effectiveness of security documentation is how quickly teams can respond to incidents. Faster response times typically reflect clear, actionable documentation that offers immediate guidance.
  • Audit success rates: Fewer audit failures or findings indicate that your documentation not only meets compliance requirements but also serves its intended operational purpose.
  • Employee feedback: Regular surveys and feedback sessions can assess whether staff find the documentation clear, accessible, and relevant to their day-to-day responsibilities. Just make sure you tailor the feedback forms even to those who aren’t big fans of excessive communication
  • Update frequency: The frequency with which documentation is reviewed and updated serves as a direct measure of its relevance and dynamism. Policies that undergo regular updates are more likely to reflect current threats and regulatory environments.

A Competitive Advantage Hiding in Plain Sight

Security documentation isn’t just about risk management—it’s about data feed management in the context of the wider security situation within the organization. 

What I mean by this is that it’s a unique opportunity to kickstart a company-wide conversation. That makes it significantly easier to unearth mistakes, misconceptions and any red flags that could be potentially calamitous if unmitigated. 

In industries where customer trust is paramount, demonstrating a mature, proactive approach to security can differentiate your brand from competitors. Strategic documentation sends a clear message: your organization takes cybersecurity seriously and is prepared to adapt to evolving threats.

Moreover, well-documented procedures can expedite certifications like ISO 27001, enhance vendor relationships, and improve the overall efficiency of your security operations.

Closing Thoughts 

It’s time to stop viewing security documentation as a necessary evil and start leveraging it as a strategic asset. 

If you embed documentation into your organization’s culture and align it with business goals, you can unlock operational efficiencies, foster resilience, and even gain a competitive edge.

The next time you’re reviewing a security policy or updating a procedure, don’t ask, ‘Does this meet compliance?’ Instead, ask, ‘Does this help us grow stronger, faster, and more secure?’ That’s the mindset shift that turns documentation from a passive obligation into a powerful, strategic tool.

Alex Williams

Alex Williams is a seasoned full-stack developer and the former owner of Hosting Data U.K. After graduating from the University of London with a Master’s Degree in IT, Alex worked as a developer, leading various projects for clients from all over the world for almost 10 years. He recently switched to being an independent IT consultant and started his technical copywriting career.

]]>
https://cacm.acm.org/blogcacm/beyond-compliance-security-documentation-as-a-strategic-asset/feed/ 0 765812
Preprinting in AI Ethics: Toward a Set of Community Guidelines https://cacm.acm.org/research/preprinting-in-ai-ethics-toward-a-set-of-community-guidelines/ https://cacm.acm.org/research/preprinting-in-ai-ethics-toward-a-set-of-community-guidelines/#respond Mon, 10 Mar 2025 15:46:53 +0000 https://cacm.acm.org/?post_type=digital-library&p=765843 Preprinting allows for the rapid dissemination of new ideas, but also of junk science and potentially of research without due ethics approval.

]]>

The fast-moving, dynamic world of artificial intelligence (AI) stands in stark contrast to the slow-moving, conservative world of academia.11 This is particularly clear in the world of AI ethics, where in addition to the industry-academia contrast we also have the meeting of very different academic disciplines, including computer science, philosophy, ethics, and social sciences. The traditions, norms, and values of these disciplines are often at odds with one another, making interdisciplinarity challenging. Take, for example, preprinting, the practice of quickly disseminating research before potentially—but not necessarily—seeking publication in traditional academic journals.a Interdisciplinary conflicts appear when, for example, researchers from a computer science background, where rapid publication of preprints on servers such as arXiv is the norm,2 meet researchers from the social sciences and humanities, where this is less common.1,30

AI ethics emerged as an active research field over the past decade, but it has a long history and a strong overlap with other fields of applied ethics, such as computer ethics.25 AI as a technology is not new either,23 nor are most of the issues with which it grapples, such as privacy, bias, and programmers’ responsibilities. The general scientific community has been portrayed as a “republic of science” by Michael Polanyi;21 some, though, have noted that this republic consists of partly overlapping domains—“fiefdoms of expertise.”10 The norms of disciplines vary greatly,1,30 and a struggle to identify the community norms around preprinting in AI ethics is currently developing, as different cultures clash in this border fiefdom. 

In this article I focus on the practice of preprinting, including the role and status of preprinting compared with traditional journal publication and other forms of dissemination, such as blogging and mainstream media contributions. Preprinting has many advantages, such as rapid dissemination of and feedback on ideas, openness, the bypassing of quarrelsome gatekeepers in the traditional publishing system, and the opportunity to publish novel ideas in new formats.4,7,27,35 In particular, it is often hailed as a beneficial practice for early-career researchers in need of quickly establishing a track record.2,19

But before we accept the idea that preprinting has many benefits and should be adopted by all, certain challenges should be noted. These include the dangers of bad science, misdirection of policy and practice, disinformation, and information overload.12,14,15,36 One important issue concerns how the growing prominence and legitimacy of preprints could more easily allow actors to use preprints in strategic efforts to influence and shape various ideals, norms, or systems. This could, for example, happen through the spread of fake science and manufactured scientific support for conspiracy theories. In addition, the seeming “democratization” of authority through preprinting can be deceptive,12 as I will argue that departing from traditional norms of scientific publishing with peer review might in fact be most beneficial for those who already wield power.

I begin with a presentation of the benefits and challenges of preprinting based on a review of the literature, including some thoughts on how these apply to AI ethics specifically. Then I suggest some broad best practices around preprinting with the aim of sparking discussion within the community.

Why Preprint?

There are many good reasons to preprint, and a quick review of the literature on preprinting in different disciplines shows clear similarities. In the following, I highlight some of the main benefits highlighted in the literature, followed by a brief discussion of how AI ethics researchers specifically might benefit from preprinting. The benefits are listed according to how often and how strongly they are reflected in the literature, with the most strongly emphasized ones discussed first. Here, I largely exclude the potential negative effects and costs related to preprinting, as these will be discussed in detail in the next section.

Speed and effective dissemination and feedback.  A primary benefit of preprinting is that it allows for the rapid and effective dissemination of new ideas.4,7,9,12,27,30,36 Many blame the publishing system and the time-consuming process of peer review for the need to preprint to get one’s ideas out there more quickly.9,12,27,32 The challenges of lengthy peer review and low acceptance rates are more severe in the humanities than in other disciplines, some argue.32 Although peer review certainly has benefits, the peer-review system has been thoroughly criticized in the broader literature, where it is seen as essentially flawed, inconsistent, slow, and ineffective.13 Preprinting is seen by some as a partial solution, or at least a way to bypass this system. Rapid dissemination can also help researchers “maintain enthusiasm” for their work,15 as the traditional review process is so slow that it can make the work feel distant when revision and potential publication occur long after the ideas were generated and written about.

Scientific deliberation and wide review.  Another major benefit pertains to how preprints can promote broad scientific deliberation, in which entire communities engage in reviews and examinations of early-stage research.2,4,7,9,14,27,36 Peer review often entails vetting by a very limited number of scholars, while preprints could, in theory, allow for vetting by a much larger audience.2 Peer review is often assumed to improve the quality of research and help identify good science,8,13 so one might expect that broader review processes could help improve science and foster researcher learning.27 The broad and rapid review that potentially ensues from preprints can also help quickly evaluate controversial results.4 For example, in 2023, a group of researchers released a preprint presenting a room-temperature superconductor17—something that, if real, would have been a huge leap in science—generating massive activity and a number of replication efforts. This was soon followed by other preprints that seemed to have debunked the original paper.20 This all took place within the span of a couple of weeks. While these sorts of events might be assumed to promote an interest in science and generate awareness, it could also be argued that this preprint, and the hype surrounding it, was wasteful and unfortunate, and that a proper vetting by peers before publication would have prevented the whole situation.

The review process for preprints varies, and includes, for example, social media discussions, counter-preprints, or straightforward peer-review reports. The latter can be found in a range of services aimed primarily at the life sciences, such as eLife, Peer Community In, and Review Commons, but the mechanics of peer review could be developed equally well for AI-ethics-relevant servers. For example, PREreview, a preprint review service, also provides integration with arXiv.

Though broad review is a potential benefit of preprints, it must be noted that just posting a preprint comes with no guarantee that it will be widely reviewed or reacted to. Though some preprints get a lot of attention, the vast majority are likely not vetted in any meaningful sense.

Author empowerment and early-career researchers.  Various stakeholders, such as researchers, funders, journals, and universities, can have different interests in and evaluations of research.2 Researchers are often touted as major beneficiaries of preprinting, with the literature emphasizing various forms of author empowerment.33 One benefit is that preprinting allows for self-archiving and both the documentation and dissemination of one’s work.1 It also potentially helps authors retain the rights to their research and control its use.19

While researchers in general could see personal benefits from preprinting, the literature stresses the benefits for early-career researchers.2,7,9,19,27 Preprints can be used in resumes and job and grant applications, and can help papers get early citations.9,19 Furthermore, they help with visibility, networking,27 and catalyzing collaboration.15 In short, preprinting increases exposure to both scholarly and popular audiences.29

Bypassing gatekeepers and the spread of novel work.  Science could be argued to symbolize and promote liberal-democratic ideals, but it simultaneously challenges the same “through its exclusivity and elitism.”3 Though there are many different academic journals and outlets in which to publish, some argue that formal publication and peer review serve a problematic gatekeeping purpose that disqualifies novel work and ideas, null results, and certain formats and styles of research. They can also be a significant barrier to groundbreaking interdisciplinary work that will suffer when evaluated by specialists in established journals.10 Preprinting can therefore help researchers bypass gatekeepers and publish novel ideas, in both traditional and novel formats.4,14,35 Vuong,35 for example, offers several reasons to preprint, based on his own experience. One was that he and his coauthors disagreed with the reviewers and editors, deciding to preprint rather than compromise by revising the paper according to reviewer recommendations. Though gatekeeping and revision after peer review might improve the quality of research, they can also remove some of the original thinking found in preprints.35

Open science and promoting access to knowledge.  The notion of open science is increasingly prevalent in the research community. Preprinting can contribute to this by promoting access to, and potentially improving the quality of, science.19,27,30 Assuming that increased access allows more actual and potential researchers to join in scientific endeavors, one might expect an increase in quality. One might also, however, imagine a situation in which openness changes and possibly undermines community standards and quality assurance procedures, which could be detrimental to the quality of science.

Related to the costs and benefits of bypassing gatekeepers, preprinting can also help “democratize” science and scientific authority.12 Openness and free publication allow researchers to bypass the structural injustices prevalent in the publishing system, where the ability and willingness to pay to publish, for example, prevail. This system gives researchers from well-off institutions significant visibility benefits. It also entails that only one aspect of the open access ideal—access to published research—is achieved, while access to open publication is reserved for the privileged. Preprinting potentially eliminates this challenge and places everyone on more equal footing.

Documenting precedence, priority, and process.  Finally, preprinting provides a public and transparent foundation of precedence and priority in research by documenting the discovery and research processes.2,4,7 It provides a transparent record of a researcher’s work,4 and “a fair and straightforward way to establish precedence.”7 It also helps document the research process from early to final-stage research, and provides insight into and proof of the co-creation process that occurs when reviewers read and comment on the various revisions preceding the publication of research.2,16

The benefits as they pertain to AI ethics.  All of the benefits of preprinting apply to the AI ethics discipline, but it is likely that technically focused researchers derive greater benefits than the more philosophically oriented members of the field.32 The need for speed—for both author and audience—is arguably greater when research is about new methods, algorithms, and so on than when it focuses on, for example, new ways of understanding the differences between and applications of ethical frameworks such as consequentialism and deontology. But social science and humanities scholars also provide vital analyses of the implications of new technologies—that is, the need for speed might be just as important for a philosopher as it is for a computer scientist.

For individual researchers, however, the benefits related to rapid dissemination, greater visibility, and a track record for applications, citations, and the like will apply regardless of whether the audience needs their research. This could lead to a situation where philosophically oriented researchers working in the same field as more technically oriented researchers start adopting the same preprint practices to stay competitive—they are, after all, competing in roughly the same job market. This is not to disregard the fact that the top journal publications are still valued in many instances, meaning that the issue is more about balance than it is an either-or question.

But if we agree that a joint or at least similar approach by all working in the same field would be beneficial, whose norms and practices should be adopted? One group might unilaterally adopt the other group’s norms. Or we could seek a compromise where both sides adjust, adopting a practice that makes the best use of preprinting while seeking traditional journals or other outlets for work best suited to those venues.

A major benefit for AI ethics is how preprinting allows researchers to overcome the difficulties that come with working in a new and interdisciplinary field. There are few specialized, high-ranking outlets where AI ethics research gets peer reviewed by traditional specialists, so this provides a good reason to preprint. It also underscores the need for new journals and outlets. This is particularly important in the field of AI, as the technology has been unleashed on societies, leaving many stakeholders—including policymakers—scrambling to find good responses. AI ethics could inform their efforts, but this would be more effective if the community had common practices for disseminating its work.

The Challenges of Preprinting in AI Ethics

Preprinting provides a wide range of potential benefits, but what does the flip side of the preprinting coin look like? Here, I summarize some of the challenges emphasized in the literature, while adding a couple of potential objections not particularly well covered in extant research.

Junk science let loose.  An obvious challenge presented by a lack of peer review is that it allows for the dissemination of poor-quality research.14,15 While erroneous and misleading results can be published quickly, they will also stay published, despite subsequent reviews and refutations. A recent example is the preprint suggesting that ChatGPT suddenly performed more poorly—and at times significantly more poorly—over time.6 The paper received a lot of attention at the time of publication, including coverage in major news outlets, despite other researchers pointing out major problems with its claims, such as its failure to note that code performance increased if one removed new surrounding code.b

That preprints can be influential is quite obvious. For example, a paper by researchers from Microsoft—an investor in OpenAI—found “sparks” of artificial general intelligence in OpenAI’s GPT-4.5 At the time of this writing (January 2024), the paper has 1,183 citations—quite good for a preprint from April 2023.

Journals, however, are subject to the same challenges. Peer review is practiced very differently across fields and publications. And there is a proliferation of new journals—including predatory ones—that offer no or only superficial peer review, yet portray published research as if it had been through rigorous peer review.

Misdirection of policy and practice.  Early results might be instrumental in responding rapidly to new situations, such as COVID-19.36 However, we have also seen that preprints can be of poor quality and misleading. When such research gets the wrong people’s attention, it could lead to misdirection of policy and practice12,15,36 and a general “confusion and distortion” of the processes that lead to proper, evidence-based policy responses to social problems.28 In the case of COVID-19, for example, a plethora of preprints were released about the methods and likelihood of transmission, and how to treat and prevent the spread of the virus.36 One suggested that the drug Ivermectin might be effective in combatting its effects, leading to the drug being administered despite a lack of evidence, with subsequent findings showing it was harmful and not effective.36

Preprints could lead some to, for example, base policies on early or uncertain results; however, it will at times be a question of some uncertain knowledge vs. no knowledge. In this case, preprints, when evaluated and used with caution, could play an important role.

Information overload.  While there are quite a few journals out there—more than most of us are able to keep track of—preprints bypassing journals altogether could exacerbate information overload.12,14 Researchers might struggle to find the most relevant research, as might the general public. More importantly, and as Sheldon28 argues, the primary curator of important science today is the media, which mainly depends on established outlets and their publication and dissemination procedures to fulfill this function. Moving to preprints, he argues, risks highlighting bad science and neglecting other important science. Though this concerns media practices, the research community must actively work with and educate the media on what the good channels are, and how to distinguish solid research from that of more uncertain quality. Though the proliferation of journals could create certain challenges for the traditional publication model, some journals will have strong, long-established reputations that they can use as a stamp of approval. To avoid complete information overload, it might be important to communicate these reputations to stakeholders beyond academia.

Some argue30 that researchers and evaluators of science are already faced with more research than they can digest. As a result, there is fear that increased use of preprints will become a time sink for researchers and grant reviewers—due to both the volume of research and the extra need for caution and careful evaluation of preprints.2

Disinformation and conspiracy theories.  Distinguishing pseudoscience from science has long been a concern in the scientific community, with some seeing the spread of pseudoscience through non-academic channels as a source of confusion for the general public.11 As such, preprints might also be seen as a new and particularly effective source of confusion. Journals arguably serve a social function by having editors and reviewers who vet research. This model is problematic and under pressure, yes, but it is still far better than nothing. When this control is bypassed, we are more vulnerable to disinformation in the form of science-like research.12,14

This could take place accidentally through the spread of junk science, but it could also occur through strategic efforts aimed at promoting or undermining various ideals, norms, or systems. While this has not received attention in the literature, science could be subject to the same type of dynamics found in social media, where “troll farms,” for example, are mobilized for political purposes.37

There is also the potential for the dissemination of articles that support or give rise to conspiracy theories,12,36 which did, in fact, occur during the COVID-19 pandemic when a preprint arguing for the “uncanny” similarities between COVID-19 and human immunodeficiency virus spurred conspiracy theories.36 These challenges will become increasingly pressing as those wanting to promote disinformation and conspiracy theories make full use of generative AI.26 Generative AI can easily be used to construct misleading or outright fake articles—even with fake data—that will likely pass most preprint servers’ relatively low bar of quality control. (Though it could also pass the bar of certain journals, I’ll argue that the probability of such articles being identified and rejected will generally be somewhat higher.)

Scooping and problems with subsequent publication.  An often-mentioned challenge in the literature is the risk of being scooped.4,14,15,27 However, most papers mention this researcher fear as being largely unfounded, arguing that preprints do not make one susceptible to scooping due to the paper trail and public nature of documented results.4 A related concern is the fear of losing out on publication opportunities if one preprints1,9,12,14,27 This challenge, though, has been mitigated by drastic changes in publisher policies, as most now allow some form of preprinting.4,7

Challenges related to power and authority.  Open science leads to more publications, and preprinting is, as we have seen, perceived as a good strategy for researchers.19 One objection to this perspective, however, is that the radical equality in the preprint world makes the playing field less even and less welcoming for unknown and early-career researchers. The idea behind blind peer review is that name and face should not matter—that, for example, Alan Turing and a rookie researcher would be equally positioned in the struggle to publish in a top journal and benefit from the “heuristic cues of a journal’s reputation, selection, and peer-review processes.”30 Much research shows that anonymous peer review is far from perfect, of course,13 but might still be better than a situation in which one’s name and reputation are immediately obvious and just about the only cue readers have to go by. Preprinting, one might argue, favors the strong—not the weak.

There are also different sources and domains of power in AI ethics. While the traditional and institutional university sector could be seen as providing one source of authority, in the world of AI ethics, a social media presence and recognition by peers are arguably just as important. Those marginalized from traditional institutions might perceive themselves as outsiders, but with a large number of social media followers, they inevitably wield power and must be considered to be powerful actors in the fiefdom of AI ethics. Social media provides a new form of power in the research community, and X/Twitter mentions, for example, are linked to the number of citations and arXiv downloads.29,36

Another way of using preprints to further power and influence is through unblinding journal reviews. If someone like Luciano Floridi, for example, submits a previously preprinted article for review in a publication, any theoretical notion of the journal’s blind review standards will evaporate. One might argue that reviewers and editors are sufficiently professional to disregard name recognition; however, this position seems naive. Grove provides a relevant anecdote about the workings of power in the academic system:

When Lord Rayleigh, already a highly reputed scientist, submitted a paper to the British Association for the Advancement of Science in 1886, his name was somehow omitted and the paper was rejected. When its authorship was discovered, it was at once accepted. 11 

Preprinting unveils power, and one might argue that blind reviews provide an imperfect, yet not wholly ineffective, veil that primarily helps those without strongly established reputations. Earlier, preprinting was mentioned as a force of democratization;12 these considerations indicate that it could also have elitist and undemocratic effects.

Research ethics.  Preprinting has the additional downside of potentially allowing research without due ethics approval, conflict of interest statements, and so on—routinely required by publishers and journals—to be disseminated.9 Bypassing such routines might, of course, be a problem because it leads to bad science, but it could also promote a situation in which researchers take greater risks, and where individuals become more exposed to both unethical research practices and potential privacy breaches following preprinting. This will largely not be the case when researchers publish preprints with an eye toward subsequent journal publication, as this will require properly dealing with ethics committees and approval in the early stages of the research, as well as in the final journal article. However, if, over time, preprints become legitimized and more accepted, it could allow those not mandated by their institutions to get comprehensive ethical clearance to both conduct and publish such research. In the traditional system, industry actors not bound by the same rules and guidance as, for example, university researchers, would be incentivized to follow traditional research ethics standards to be able to publish in top journals.

Manipulation of metrics and sector incentives.  Finally, a minor objection concerns how preprints can be used to boost a researcher’s publication count and citation score. For example, preprints are indexed by Google Scholar, thus opening the possibility of inflating both article and citation counts through prolific preprinting. Preprints can therefore be used to game and exploit the various indicators increasingly prevalent in the world of research.

The challenges as they pertain to AI ethics.  The negative aspects of preprinting all apply to AI ethicists, but to varying degrees, depending on the topic and type of research.

First, there is a lot of hype around AI; deployment of AI systems is moving very rapidly, and policymakers and other stakeholders are scrambling to identify the best responses. This suggests that the speed and flexibility of preprinting could be particularly valuable in this field, but it also shows that the potential for harm is significant. Low-quality research will always be detrimental to the scientific community, but it is arguably more harmful when it is in danger of being misinterpreted or strategically misused to inform policy or practice in harmful ways. In addition, low-quality research contributes to the crowding of the marketplace of ideas and risks increasing information overload and burying high-quality research.

Second, the practice of acknowledging and citing preprints could promote the acceptance of unethical research conducted without sufficient safeguards for either society or individuals. It is therefore incumbent upon the community as a whole to be restrictive when citing and using research that does not adhere to the ethical standards one strives for, and this requires careful attention to data availability, protocols, statements, and so on.

Third, the community should be aware of the potential for powerful actors and groups to bypass community institutions and change the norms of the fiefdom by first gaining status and then simply preprinting, without an eye to publication. This would undermine whatever quality control we have in the existing faulty system, and could also lead to a situation where preprinting becomes standard practice, as others start citing and using preprints actively in their own research.

Finally, as preprints can be abused for disinformation and other manipulation efforts, it is important that the community stay alert to such efforts to influence either the scientific community or the public through dissemination of propaganda masked as scientific preprints.

Toward Best-Practice Preprinting in AI Ethics

The preceding sections have detailed a range of benefits and challenges related to preprinting, summarized in Table 1.

Table 1.  Summary of the benefits and challenges of preprinting.
Aspect Benefits of Preprinting Challenges of Preprinting
Speed and Efficiency
  • Rapid dissemination of findings

  • Early feedback and engagement from the community

  • Risk of spreading preliminary results that may be flawed

  • Potential for rapid dissemination of bad science

Accessibility
  • Enhances visibility of research, especially for early-career researchers

  • Bypasses traditional gatekeepers

  • Can lead to information overload

  • Difficulties in navigating the increased volume of literature

Review Process
  • Broader scientific deliberation and review

  • Engages a wider audience in the critique process

  • Lack of formal peer review can undermine quality control

  • High variability in reviewer expertise

Empowerment
  • Authors maintain control over their work

  • Helps establish track records and boost citations

May disproportionately benefit established researchers with wider recognition and networks
Openness
  • Promotes open science and transparency

  • Facilitates collaboration and sharing of ideas

  • Challenges in maintaining quality standards

  • Potential misuse of preprints for promoting pseudoscience

At a time when disinformation,37 polarization,31 and a lack of trust in the scientific community in large groups18 are prevalent, it is important that the scientific community considers its role and responsibilities and actively engages in the shaping of our academic culture and practices. This is no less relevant for the field of AI ethics, where we deal with technologies that are rapidly implemented, with major implications for individuals, businesses, and governments. As mentioned, there is much hype concerning AI, and a high demand for AI ethics expertise. There is a great temptation to jump into the field, and to exploit opportunities that bring recognition from the community.24 When things move fast, the traditional world of scientific publishing feels exceedingly slow. As a journal editor myself, I see firsthand how researchers might struggle as their studies of brand-new technologies are stuck in review and revision processes that mean their results will be outdated by the time they are finally published. However, I would argue that any manuscript that would actually be outdated because, for example, a software solution has been updated, is not appropriate for publication anyway, as it should contain analyses of the solution and considerations related to what could be learned and developed based on the results of the research. These should be valid and useful regardless of how rapidly the technology develops. Manuscripts that do not have these more timeless elements are perhaps better suited for other forms of dissemination, such as Medium or Substack posts. Or preprints?

One key difference between preprints and these other forms of publishing, however, is that preprints are more likely to stay available, accessible, and usable by other researchers than alternative formats. Preprints get a persistent identifier such as a DOI, and are indexed in scientific search engines. Another user benefit is that preprints cannot be changed without providing a record of changes. All of this makes the preprint a more solid form of dissemination likely to be used and referenced by other researchers, potentially helping promote the broader review and extensive discussion that I’ve discussed here.

In the following, I propose a tentative set of guidelines for preprinting, intended as a basis for community discussion and debate.

Consider alternative formats.  Before preprinting, consider how to best disseminate your ideas to the relevant stakeholders. A key determinant of whether to preprint relates to the scholarly nature and ambition of the work. Only work that aspires to be scholarly or scientific research should be preprinted; this helps avoid the further dilution of preprinting as a valuable scientific practice. Much solid work is better suited for publication in mainstream or technology-specific magazines, blogs, or other outlets. Considering these alternative formats is important to reduce one’s contribution to information overload in scientific channels.

Consider the need for speed.  Are your ideas of a kind that will be immediately useful, and potentially even important, for stakeholders in the scientific community? Or is it important for you to establish precedence with your research—potentially without the need for broad dissemination? If the answer to either of these questions is yes, and if your work adheres to the relevant expected scientific standards, preprinting could be a good idea. If not, consider the downside of too rapid publication. The consequences of premature dissemination of research that does not need rapid publication could be that a) the ideas lose novelty, b) the ideas are discredited due to the immature state of the research, and, consequently, c) your reputation suffers.

Consider the need for broad review.  If your ideas require a particularly extensive process of review and discussion, preprinting might allow for a broad “review” process before you finalize the research. However, this can also be achieved by using blogs or other outlets, as demonstrated by John Danaher, who uses his blog to publish partial and tentative analyses of his work.c Some might also use mainstream media for testing and getting feedback on new ideas. This naturally requires that the research have a sufficiently broad appeal, or that researchers have access to journalists or others interested in aiding such efforts. If you preprint for broad review, make sure to state this clearly in the preprint, and spread it with an explicit request for feedback. Otherwise, one of the key benefits of preprinting is lost. As preprints are now used for so many different purposes, most of your peers will not realize that, just by posting a paper on a preprint server, you are looking for comments and feedback.

Consider the potential positive and negative consequences.  How can and will the preprint be used? If you see yourself as a free and isolated scientist with no responsibility for how others interpret and use your findings, this question might be irrelevant. Otherwise, you should know that your research will always have social implications, so be sure to consider the potential for your non-reviewed research to misdirect policy or practice. In the fields of AI and AI ethics, it will be particularly relevant to consider how your work might either contribute to AI hype or lead some to not adopt technologies that are societally beneficial. Such effects might be desired, but they could have major consequences and should ideally be the result of solid research that is not prematurely published.

Adhere to academic standards of research ethics when possible.  AI is a field in which academics are by no means the only participants, or even the major ones. While researchers in universities and, for example, the healthcare sector tend to be subject to relatively strict standards of research ethics, this does not apply to AI ethicists working in tech companies. Researchers from companies such as Meta, Microsoft, and OpenAI can do things other researchers cannot, and though much industry research would be stopped at the gates by journals requiring adherence to ethical standards, these organizations are free to preprint seemingly scholarly work without any scholarly standards of ethics. This is a good argument for academic researchers to seek journal publication; however, it could also be a reason for industry researchers to voluntarily adhere to standards of ethics, to help their research gain broader acceptance and legitimacy. I argue that there are reasons for the establishment of research ethics standards and that doing so might be good regardless, and perhaps should even be mandated more broadly.

If you preprint: Engage, revise, and redact when appropriate.  Many of the benefits of preprinting require you to actively disseminate your research and engage with those who read and comment on it. Ideally, this can and should lead to revisions of the work. And in instances where your audience discovers grave errors in your preprint, it should be updated and, if possible, redacted. Without such engagement, preprinting is just a self-archival and self-promotion tool—most of the benefits are lost.

Consider your own role and authority.  For established researchers who no longer feel they need to prove themselves through cumbersome peer review processes, it can be tempting to post everything as preprints. However, even the best among us might need some correction, and even if the publish-everything strategy might be successful in gaining attention and citations, raising oneself above one’s community risks both hurting the community and, over time, undermining oneself. A related consideration is the disproportionate power wielded by authorities in the field to both raise up and hurt others. Considering impacts on others is important, especially for early-career researchers just starting out. Preprinting willy-nilly exacerbates the risks of arbitrarily using social community power.

Conclusion

A diverse field such as AI ethics is bound to have a number of fault lines separating various groups with different backgrounds. This, though, does not necessarily mean that the various groups are in conflict, or that they cannot learn from each other and cooperate.10 Bridging the fault lines, rather than shouting across the chasm without any real attempt to understand or constructively learn from and engage with each other, should be preferable. Fostering such cooperation, however, entails actively engaging in and developing community guidelines, coordination, and a joint understanding of these different backgrounds and perspectives.38 This discussion of preprinting practices is an attempt to chip away at one small part of this challenge.

Preprinting can be highly beneficial—especially in AI ethics—but would benefit from adherence to certain guidelines to avoid some of the pitfalls. Even when these—or other, better—guidelines are followed, additional questions remain.

One important issue is how we can help each other navigate and leverage a rising tide of research of increasingly unknown merit. There is no universally accepted definition of what constitutes scientific knowledge,22 but it is commonly identified through its recognition and appropriation by the scientific community.8,11 Developing venues for community debate and deliberation will be important to effectively make use of and curate preprints. It might also make sense to find, shape, and develop more interdisciplinary preprint servers for AI-ethics-related work, as existing servers are poorly suited for this purpose. Furthermore, these debates must include other stakeholders, including the media and citizens. Navigating research in the field is difficult enough for those in it, so making sure we help others understand what research can be trusted—for what purposes and how much—will be crucial as preprinting practices develop.

]]>
https://cacm.acm.org/research/preprinting-in-ai-ethics-toward-a-set-of-community-guidelines/feed/ 0 765843
Barto, Sutton Announced as ACM 2024 A.M. Turing Award Recipients https://cacm.acm.org/news/barto-sutton-announced-as-acm-2024-a-m-turing-award-recipients/ https://cacm.acm.org/news/barto-sutton-announced-as-acm-2024-a-m-turing-award-recipients/#respond Wed, 05 Mar 2025 15:15:58 +0000 https://cacm.acm.org/?p=766007 ACM has named Andrew G. Barto and Richard S. Sutton as the recipients of the 2024 ACM A.M. Turing Award for developing the conceptual and algorithmic foundations of reinforcement learning.

In a series of papers beginning in the 1980s, Barto and Sutton introduced the main ideas, constructed the mathematical foundations, and developed important algorithms for reinforcement learning—one of the most important approaches for creating intelligent systems.

Their work laid the groundwork for the practical application of reinforcement learning in its merging with deep learning.

Barto is Professor Emeritus of information and computer sciences at the University of Massachusetts, Amherst. Sutton is a professor of computer science at the University of Alberta, a Research Scientist at Keen Technologies, and a Fellow at Alberta Machine Intelligence Institute.

Although Barto and Sutton’s algorithms were developed decades ago, major advances in the practical applications of RL came about in the past fifteen years by merging RL with deep learning algorithms

Read More

]]>
https://cacm.acm.org/news/barto-sutton-announced-as-acm-2024-a-m-turing-award-recipients/feed/ 0 766007
Program Merge: What’s Deep Learning Got to Do with It? https://cacm.acm.org/practice/program-merge-whats-deep-learning-got-to-do-with-it/ https://cacm.acm.org/practice/program-merge-whats-deep-learning-got-to-do-with-it/#respond Thu, 20 Feb 2025 15:48:59 +0000 https://cacm.acm.org/?post_type=digital-library&p=764866 Leading figures of Microsoft Research's DeepMerge project discuss their efforts to apply machine learning to complicated program merges.

]]>

If you regularly work with open-source code or produce software for a large organization, you are already familiar with many of the challenges posed by collaborative programming at scale. Some of the most vexing of these tend to surface as a consequence of the many independent alterations inevitably made to code, which, unsurprisingly, can lead to updates that do not synchronize.

Difficult merges are nothing new, of course, but the scale of the problem has gotten much worse. This is what led a group of researchers at Microsoft Research (MSR) to take on the task of complicated merges as a grand program-repair challenge—one they believed might be addressed at least in part by machine learning (ML).

To understand the thinking that led to this effort and then follow where that led, ACM Queue asked Erik Meijer and Terry Coatta to speak with three of the leading figures in the MSR research effort, called DeepMerge.a Meijer was long a member of MSR, but at the time of this discussion was director of engineering at Meta. Coatta is the chief technology officer of Marine Learning Systems. Shuvendu Lahiri and Christian Bird, two of the researchers who helped drive this effort, represent MSR, as does Alexey Svyatkovskiy, who was with Microsoft DevDiv (Development Division) at the time.

Terry Coatta: What inspired you to focus on merge conflicts in the first place? And what made you think you’d be able to gain some advantage by applying AI techniques?

Christian Bird: Back in the winter of 2020, some of us started talking about ways in which we might be able to use machine learning to improve the state of software engineering. We certainly thought the time was right to jump into an effort along these lines in hopes of gaining enough competency to launch into a related research program.

We tried to identify problems other researchers weren’t already addressing, meaning that something like code completion—which people had been working on for quite some time—was soon dismissed. Instead, we turned to problems where developers didn’t already have much help.

Shuvendu [Lahiri] has a long history of looking at program merge from a symbolic perspective, whereas my own focus has had more to do with understanding the changes that occur in the course of program merges. As we were talking about this, it dawned on us that almost no one seemed to be working on program merge. And yet, that’s a problem where we, as developers, still have little to rely upon. For the most part, we just look at the diffs between different generations of code to see if we can figure out exactly what’s going on. But there just isn’t much current tooling to help beyond that, which can prove to be problematic whenever there’s a merge conflict to resolve.

So, we figured, “OK, let’s look at how some deep-learning models might be applied to this problem. As we go along, we’ll probably also identify some other things we can do to build on that.”

Shuvendu Lahiri: Yes, as Chris suggests, I’ve been thinking about the issues here for quite some time. Moreover, we found program merge to be appealing, since it’s a collaboration problem. That is, even if two skilled developers make correct changes, the merge itself may introduce a bug.

We were also keenly aware of the sort of pain program-merge problems can cause, having known about it through studies within Microsoft.b I thought maybe there was something we could do to provide relief. It also turns out that AI was just coming along at that point, and Alexey [Svyatkovskiy] had already developed a couple of powerful models that looked quite promising for code completion. What’s more, information about merge conflicts was just starting to become more readily available from the Git commit history, so that too looked like it might serve as a good up-front source of clean data.

Erik Meijer: I like the fact that you focused on merge conflict, since, when it comes to this, I don’t think source control solves any of the real problems. Maybe I’m being a little extreme here, but even if source control lets you know where you have a merge conflict, it won’t help you when it comes to actually resolving the conflict. In fact, I’m baffled as to why this problem wasn’t solved in an intelligent manner a long time ago. Are people just not listening to complaints from actual users?

Lahiri: Basically, I think it comes down to academicians consistently resorting to symbolic methods to solve this problem. Whereas people in the real world have looked at this as just another aspect of programming, practitioners have been more inclined to approach it as a social process—that is, as a problem best addressed by encouraging co-workers to figure out solutions together. Personally, I’ve always seen merge conflicts as more of a tooling challenge.

Alexey Svyatkovskiy: For me, this just looked like an exciting software engineering problem to address with machine learning. I’ve spent years working on code completion, but this effort looked like something that would take that up to the next level of complexity, since it necessarily would involve aligning multiple sequences somehow and then complementing that with an understanding of where things ought to be inserted, deleted, or swapped. And, of course, there were also those special cases where the developer would be able to add new tokens during the merge.

This took us on a journey where we ended up addressing program merge down at the line-and-token level. I found this fascinating, since a lot of people don’t have any idea about how merge actually works and so, by extension, don’t have a clear understanding about what leads to merge conflicts. Taking on this problem also seemed important in that, while merge conflicts are far less common than software bugs, they require considerably more time to resolve and can end up causing more pain.

Coatta: How did you initially attack the problem?

Lahiri: We realized that repositories (both open-source ones on GitHub and internal ones at Microsoft) contain data on merge conflicts and their resolution across several different programming languages. What’s more, Alexey had recently created a neural model that had been pre-trained on a large subset of the code in those repositories. Our initial thought was to fine-tune that model with information about merge conflicts and their resolution. And we figured that should be simple enough to be treated as an intern project. So, that’s how we scoped it initially. We figured: Just get the data, train a model, and deploy. What we failed to grasp was that, while there was an ample amount of program merge data to be mined, coming to an understanding of what the intent behind all those merges had been was at least as important as the data itself. In fact, it proved to be absolutely critical.

A considerable amount of time and effort was required to understand and interpret the data. This included some significant technical challenges—for example, how best to align these programs. And how can you communicate to a neural model that these are not independent programs but instead represent some number of changes to an underlying program? The notion of how to go from program text to a program edit became quite crucial and, in fact, required considerable research. Ultimately, we concluded that if you manage to combine your existing merge tools correctly, add just the right amount of granularity—which for us proved to be tokens—and then employ neural modeling, you can often succeed in reducing the complexity. But it took us quite a bit of time to work that out.

Of course, we also underestimated the importance of user experience. How exactly would a user end up employing such a tool—one that’s AI-based, that is? And what would be the right time to surface that aspect of the tool?

Coatta: I find it fascinating that it proved to be so difficult to scope this project correctly. Can you dig a bit deeper into that?

Svyatkovskiy: To me, at least, as we were analyzing the different types of merges, it soon became clear that there are varying levels of complexity. Sometimes we’d find ourselves looking at two simple merge resolution strategies, where it essentially came down to “Take ours or take theirs.” Such cases are trivial to analyze, of course, and developers don’t require much AI assistance when it comes to resolving these conflicts.

But then there’s another class of merge, where a new interleaving line is introduced that involves more than just concatenation. There could also be token-level interleaving, where lines in the code have been broken and new tokens introduced in between. This leads to the notoriously complex case where a switch to token-level granularity proves to be crucial. Beyond that, there’s a whole other class of merges where you find somebody has introduced some new tokens.

Meijer: How do you go about defining what you consider to be a correct merge? Doesn’t that require you to make your own value judgments in some sense?

Lahiri: Well, I’ll just say we had a very semantic way of looking at merges. Essentially: “Forget about the syntax; instead, what does it mean for the merge to be correct?” In effect, this amounts to: If something was changed in one program, then that ought to be reflected in the merge. And, if that also changes a behavior, then that too ought to be included in the merge. But no other changes or altered behaviors should be introduced.

We then found, however, that we could get tangled up whenever we ran into one of these “take my changes or take yours” merges. We also found that one set of changes would often just be dropped—like a branch being deprecated, as Alexey once pointed out. This is how we discovered that our initial notion of correctness didn’t always hold. That’s also how we came to realize we shouldn’t adhere to overly semantic notions.

So, we decided just to do our best to curate the training data by removing any indications of “take your changes or take mine” wherever possible. Then we looked at those places where both changes had been incorporated to some extent and said, “OK, so this now is our ground truth—our notion of what’s correct.” But notice that this is empirical correctness as opposed to semantic correctness—which is to say, we had to scale back from our original high ambitions for semantic correctness.

Svyatkovskiy: We now treat user resolutions retrieved from the GitHub commit histories as our ground truth. But yes, naturally, there are all kinds of ways to define a “correct” merge. For example, it’s possible to reorder the statements in a structured merge and yet still end up with a functionally equivalent resolution. Yet, that would be deemed as “incorrect,” so, there’s clearly room for retooling our definition of correctness. In this instance, however, we chose to take a data-driven approach that treats user resolutions from the GitHub commit histories as our ground truth.

Bird: Right. And let me also say that, from the beginning of this project, we decided to approach it as something that might yield a product. With that in mind, we realized it needed to be, perhaps not language-agnostic, but at least something that could be readily adapted to multiple languages—and definitely not something that would require some bespoke analysis framework for each language. That essentially guided our choice not to employ richer or more complex code representations.

Meijer: I’ve also run into situations like this where it looked really tempting to use an AST [abstract syntax tree] or something of the sort, since that would provide all the structure that was required. But then, as you go deeper into that sort of project, you find yourself wondering whether it’s actually a good idea to feed semantically rich programs into models and start thinking it might be better just to send strings instead.

Coatta: To dive a bit deeper into that, you had a practical motivation to work with a token-based approach. But what does your intuition tell you about how models behave when you do that? If you feed a model a rich, structured information set, is that model then actually more likely to make better decisions? Or is that perhaps a false assumption?

Svyatkovskiy: The model ought to be able to make better decisions, I think.

Meijer: All right, but can I challenge that a bit? The model can handle the syntax and semantic analysis internally, which suggests this work may not need to be done ahead of time, since machines don’t look at code the same way humans do. I don’t know why the model couldn’t just build its own internal representation and then let a type-checker come along at the end of the process.

Bird: I think it’s hazardous to speculate about what models may or may not be capable of. I mean, that’s a super-nuanced question in that it depends on how the model has been trained and what the architecture is like—along with any number of other things. I’m constantly surprised to learn what models are now capable of doing. And, in this case, we’re talking about the state of the world back in 2020—even as I now find it hard to remember what the state of the world looked like six months prior to the GPT models becoming widespread.

Lahiri: For one thing, we were using pretrained models to handle classification and generation, which then left us with quite a bit of work to do in terms of representing the resulting edits at the AST level before tuning for performance. That certainly proved to be a complex problem—and one that came along with some added computational costs. Also, as I remember it, the models we were using at the time had been trained as text representations of code—meaning we then needed to train them on a lot more AST-level representations to achieve better performance. I’m sure it would be fascinating to go back to revisit some of the decisions we made back in 2020.

Meijer: What model are you using now?

Svyatkovskiy: For this iteration, we’re employing a token-level merge along with a transformer-based classifier. We’ve also been looking at using a prompt-driven approach based on GPT-4.

Meijer: I love that this is now something where you can take advantage of demonstrated preferences to resolve merge conflicts instead of being left to rely solely on your own opinions.

Lahiri: Another way of looking at this that came up during our user studies was that, even after a merge has been produced, someone might want to know why the merge was accomplished in that particular way and may even want to see some evidence of what the reasoning was there. Well, that’s a thorny issue.

But one of the nice things about these large foundational models is that they’re able to produce textual descriptions of what they’ve done. Still, we haven’t explored this capability in depth yet, since we don’t actually have the means available to us now to evaluate the veracity of these descriptions. That will have to wait until some user studies supply us with more data. Still, I think there are some fascinating possibilities here that ultimately should enable us to reduce some of the friction that seems to surface whenever these sorts of AI power tools are used to accomplish certain critical tasks.

 

Coatta: You’ve mentioned that you had access to a vast amount of training data, but you’ve also suggested some of that data contained surprises—which is to say it proved to be both a blessing and a curse. Can you go into that a bit more?

Lahiri: Yes, we were surprised to find that a large percentage of the merges—perhaps 70%—had the attribute of choosing just one side of the edit and then dropping the other. In some of those cases, it seemed one edit was superseding the others, but it can be hard to be sure whenever the syntax changes a little. In many instances, there were genuine edits that had been dropped on the floor. It was unclear whether that was due to a tooling problem or a social issue—that is, in some cases, perhaps some senior developer’s changes had superseded those that had been made by a junior developer. Another hypothesis was that, instead of a single merge, some people may have chosen to merge in multiple commits.

This sort of thing was so common that it accounted for a significant portion of the data, leaving us uncertain at first as to whether we should throw out these instances, ignore them, or somehow make an effort to account for them. That certainly proved to be one of the bigger surprises we encountered.

Another surprise was that we discovered instances where some new tokens had been introduced that were irrelevant to the merge. It was unclear at first whether those were due to a genuine conflict in the merge or just because somebody had decided to add a pretty print statement while doing the refactoring. That proved to be another thorny issue for us.

Coatta: How did you resolve that? It sounds like you had some datasets you didn’t quite know how to interpret. So, how did you decide what should be classified as correct merges or treated as incorrect ones?

Lahiri: We curated a dataset that did not include the “trivial” merge resolutions, with the goal of assisting users with the more complex cases first. As Alexey mentioned, users may not need tooling support for those resolutions that only require dropping one of the two edits.

Svyatkovskiy: And then, from user studies, we learned that some users still wanted to be able to use the approach that had been dismissed. We solved that problem by providing a “B option” that people could get to by using a drop-down menu.

Lahiri: Which is to say we addressed the problem by way of user experience rather than by changing the model.

The other data problem we encountered had to do with new tokens that would occasionally appear. Upon closer examination, we found these tokens were typically related to existing changes. By going down to token-level merges, we were able to make many of these aspects go away. Ultimately, we built a model that excluded that part of the dataset where new tokens were introduced.

Meijer: In terms of how you went about your work, I understand one of the tools you particularly relied on was Tree-sitter [a parser-generator tool used to build syntax trees]. Can you tell us a bit about the role it played in your overall development process?

Bird: We were immediately attracted to Tree-sitter because it lets you parse just about anything you can imagine right off the shelf. And it provides a consistent API, unlike most other parsers out there that each come with their own API and work only with one language or another.

For all that, I was surprised to learn that Tree-sitter doesn’t provide a tokenizing API. As an example of why that proved to be an issue for us, we wanted to try Python, which basically lets everyone handle their own tokenizing. But, of course, Tree-sitter didn’t help there. We resorted to a Python tokenizing library.

Beyond that relatively small complaint, Tree-sitter is great in terms of letting you apply an algorithm to one language and then quickly scale that up for many other languages. In fact, between that capability and the Python tokenizing library, which made it possible for us to handle multiple languages, we were able to try out things with other languages without needing to invest a lot of upfront effort. Of course, there’s still the matter of obtaining all the data required to train the model, and that’s always a challenge. At least we didn’t need to write our own parsers, and the consistent interfaces have proved to be incredibly beneficial.

Meijer: Once you finally managed to get all this deployed, what turned out to be your biggest surprise?

Bird: There were so many surprises. One I particularly remember came up when we were trying to figure out how people would even want to view merge conflicts and diffs. At first, some of us thought they’d want to focus only on the conflict itself—that is, with a view that let them see both their side and the other side. It turns out you also need to be able to see the base to understand the different implications between an existing branch in the base and your branch.

So, we ran a Twitter survey to get a sense of how much of that people thought we should show. How much of that did they even want to see? For example, as I recall, most people couldn’t even handle the idea of a three-way diff, or at least weren’t expecting to see anything quite like that. That really blew my mind, since I don’t know how anyone could possibly expect to deterministically resolve a conflict if they don’t know exactly what they’re facing.

Some other issues also came up that UI people probably would expect, but I nevertheless was incredibly surprised. That proved to be a big challenge, since we’d been thinking throughout this whole process that we’d just get around to the UI whenever we got around to it. And yes, as this suggests, our tendency initially was just to focus on making sure the underlying algorithm worked. But then we found to our surprise just how tough it could be to find the right UI to associate with that.

Coatta: From what you say, it seems you weren’t surprised about the need for a good user experience, but it did surprise you to learn what’s considered to be a good experience. What are your thoughts now on what constitutes a good user experience for merge?

Bird: I’m not entirely clear on that even now, but I’ll be happy to share some of the things we learned about this early on. As we’ve already discussed, people definitely want to see both sides of a merge. Beyond that, we discovered that they want the ability to study the provenance of each part of the merge because they want to know where each token came from.

So, we wrote some code to track each token all the way back to whichever side it came from.

There also were tokens that had come in from both sides. To make it clear where a token had originated, we wrestled with whether we should add colors as an indicator of that. How might that also be used to indicate whether a token happens to come from both sides or simply is new?

In addition, we knew it was important that the interface didn’t just ask you to click “yes” or “no” in response to a suggested change, since it’s rare to find any merge that’s going to be 100% correct. Which is to say developers are going to want to be able to modify the code and will only end up being frustrated by any interface that denies them that opportunity.

The real challenge is that there are lots of moving pieces in any given merge. Accordingly, there are many possible views, and yet you still want to keep things simple enough to avoid overwhelming the user. That’s a real challenge. For example, we know that if we offer three suggestions for a merge rather than just one, the chance of the best one being selected is much higher. But that also adds complexity, so we ultimately decided to go with suggesting the most likely option, even though that might sometimes lead to less-optimal results.

There are some other user-experience considerations worth noting. For example, if you are working on some particular Visual Studio feature, you’re going to want to produce something that feels intuitive to someone who has been using that same tool. Suffice it to say, there’s plenty to think about in this respect. Basically, once you finally get your model to work, you might not even be halfway home, since that’s just how critical—and time-consuming—the user-experience aspect of this work can be.

Coatta: We all know that creating a tool for internal purposes is one thing, while turning that into a product is something else altogether. It seems you took that journey here, so what were some of the bigger surprises you encountered along the way?

Lahiri: Actually, we don’t have a product yet that implements the DeepMerge algorithm and aren’t at liberty to talk about how that might be used in future products. Still, as we’ve just discussed, I can say most of the unusual challenges we encountered were related to various aspects of the user experience. So, we got much deeper into that here than we normally would.

One of the biggest challenges had to do with determining how much information needed to be surfaced to convince the user that what was just done was even possible—never mind appropriate. Suddenly, you’ve just introduced some new tokens over here, along with a new parsable tree over there. I think that can really throw some users off.

Bird: What did all this look like from the DevDiv perspective, Alexey? You deal with customers all the time. What proved to be the biggest challenges there?

Svyatkovskiy: Some of the most crucial design decisions came down to choosing between client-side or server-side implementation. Our chief concern had to do with the new merge algorithm we were talking about earlier. Customer feedback obtained from user studies and early adopters proved to be particularly crucial in terms of finding ways to smooth things out. Certainly, that helped in terms of identifying areas where improvements were called for, such as achieving better symmetries between what happens when you merge A to B versus when you merge B to A.

Lahiri: I’d like to add a couple of points. One is that some developers would prefer to handle these merges themselves. They just don’t see the value of tooling when it’s used to deal with something they could do themselves. But that just resulted in some inertia, which is always hard to overcome without a lot of usage. Still, from our empirical study we learned that, even when merges were not identical to the ground truth, users would accept them if they proved to be semantically equivalent. Ultimately, that proved to be a pleasant surprise, since it revealed we had previously been undercounting our wins according to our success metrics.

Coatta: Did anything else interesting surface along the way?

Bird: At one point, one of our interns did a user study that pulled merge conflicts and their resolutions out of Microsoft’s historical repositories so they could then be compared with the resolutions our tool would have applied. As you might imagine, quite a few differences surfaced. To understand where our tool may have gone wrong, we went back to consult with those people who had been involved in the original merges and showed them a comparison between what they had done and how the tool had addressed the same merge conflicts.

We specifically focused on those conflicts that had been resolved over the preceding three months on the premise that people might still recall the reasoning behind those decisions. We learned a ton by going through that particular exercise. One of the lessons was that we had probably undercounted how often we were getting things right, since some of these developers would say things like, “Well, this may not exactly match the merge I did, but I would have accepted it anyway.”

The other major benefit of that study was the insight it provided into what the user experience for our tool should be. This all proved to be a major revelation for me, since it was the first time I’d been involved in a user study that was approached in quite this way—where developers were pulled in and presented with code they’d actually worked on.

Which is just to say this wasn’t at all like one of those lab studies where people are presented with a toy problem. In this case, we were pulling in real-world merge conflicts and then talking with the developers who had worked to resolve them. We learned so much from taking this approach that I’d recommend other researchers consider doing their own studies in much the same way.

Svyatkovskiy: Another thing that came out of these user studies was the importance of explainability. With large language models, for example, we can drill into specific three-way diffs and their proposed resolutions and then ask for summaries of certain decisions, which can be helpful when it comes to building confidence in some of these AI suggestions.

Also, as Chris indicated, even when users chose not to go with the solution offered by DeepMerge, the reasoning behind the suggestion still seemed to inform their own thinking and often led to an improved merge resolution.

Coatta: What’s next?

Lahiri: There’s room for more prompt engineering in terms of determining what goes into the model’s input. We also need to address correlated conflicts. So far, we’ve addressed each conflict as if it was independent, but you can have multiple conflicts in a file that all relate to a certain dependency. Some users have told us that, once a resolution has been worked out for one of those conflicts, they’d like to see something similar applied to each of the other conflicts that exhibit a similar pattern, which certainly seems quite reasonable.

Also, while the types of conflicts we’ve addressed so far are highly syntactic in nature, there is, in fact, a whole spectrum of merge conflicts. There’s still much to address, including silent merges that include semantic conflicts, which are much harder to deal with than anything we’ve handled so far. Still, I’d say it feels like we’re off to a reasonably good start.

From
ACM Queue logo

 

]]>
https://cacm.acm.org/practice/program-merge-whats-deep-learning-got-to-do-with-it/feed/ 0 764866
Abstractions https://cacm.acm.org/opinion/abstractions/ https://cacm.acm.org/opinion/abstractions/#comments Wed, 19 Feb 2025 17:15:13 +0000 https://cacm.acm.org/?post_type=digital-library&p=764610 Abstraction is a by-product of computer science's central purpose, understanding information processes.

]]>

We claim two things about our profession. Computer science studies information processes, natural and artificial. Computer science is a master of abstraction. To reconcile the two, we say that abstraction is the key that unlocks the complexity of designing and managing information processes.

Where did we get the idea that our field is a master of abstractions? This idea is cosmic background radiation left over from the beginning bangs of our field. For its first four decades, computer science struggled under often blistering criticisms by scientists that the new field was not a science. Science is, they said, a quest to understand nature—and computers are not natural objects. Our field’s pioneers maintained that computers were a major new phenomenon that called for a science to understand them. The most obvious benefit of our field was software that controls and manages very complex systems. This benefit accrued from organizing software into hierarchies of modules, each responsible for a small set of operations on a particular type of digital object. Abstraction became a shorthand for that design principle.

Approximately 25 years ago, the weight of opinion suddenly flipped. Computer science was welcomed at the table of science. The tipping point came because many scientists were collaborating with computer scientists to understand information processes in their fields. Many computer scientists claimed we had earned our seat because of our expertise with abstractions. In fact, we won it because computation became a new way of doing science, and many fields of science discovered they were studying naturally occurring information processes.

In what follows, I argue that abstraction is a by-product of our central purpose, understanding information processes. Our core abstraction is information process, not “abstraction.” Every field of science has a core abstraction—the focus of their concerns. In all but a few cases, the core abstractions in science defy precise definitions and scientists in the same field disagree on their meanings. Two lessons follow. First, computing is not unique in believing it is a master of abstraction. Indeed, this claim never sat well with practitioners in other fields. Math, physics, chemistry, astronomy, biology, linguistics, economics, psychology—they all claim to be masters of abstractions. The second lesson is that all fields have made remarkable advances in technology without clear definition of their core abstraction. They all designed simulations and models to harness the concrete forces behind their abstractions. The profound importance of these lessons was recognized with two 2024 Nobel prizes awarded to computer scientists for protein folding and machine learning.

What Is Abstraction?

Abstraction is a verb: to abstract is to identify the basic principles and laws of a process so that it can be studied without regard to physical implementation; the abstraction can then guide many implementations. Abstraction is also a noun: An abstraction is a mental construct that unifies a set of objects. Objects of an abstraction have their own logic of relations with each other that does not require knowledge of lower-level details. (In computer science, we call this information hiding.)

Abstractions are a power of language. In his book Sapiens, Yuval Noel Harari discusses how, over the millennia, human beings created stories (“fictions”) that united them into communities and gave them causes they were willing to fight for.4 These fictions were abstractions that often endured well beyond their invention. The U.S. constitution, for instance, applies to all its states and has guided billions of people for more than 200 years.

The ability of language to let us create new ideas and coordinate around them also empowers language constructs to refer to themselves. After all, we have numerous ideas about our ideas. We build endlessly complex structures of ideas. We can imagine things that do not exist, such as unicorns, or unrealized futures that we can pursue. Self-reference also generates paradoxes. A famous paradox asks: “Does the set of all those sets that do not contain themselves contain itself?” Self-reference is both a blessing and a curse.

Ways to avoid paradoxes are to stack up abstractions in hierarchies or connect them in networks. An abstraction can be composed of lower-level abstractions but cannot refer to higher level abstractions. In chemistry, for example, amino acids are composed of atoms, but do not depend on proteins arranging the acids in particular sequences. In computing, operating systems are considered layers of software that manage different abstractions such as processes, virtual memories, and files; each depends on lower levels, but not higher levels. Consider three examples illustrating how different fields use their abstractions.

Computer science.  An “abstract data type” represents a class of digital objects and the operations that can be performed on them. This reduces complexity because one algorithm can apply to large classes of objects. The expressions of these abstractions can be compiled into executable code: thus, abstractions can also be active computing utilities and not just descriptions.

Physics.  For physicists, abstraction simplifies complex phenomena and enables models to help understand and predict the behavior of complex systems. Many physics models take the form of differential equations that can be solved on grids by highly parallel computers. For example, the Stokes Equation in computational fluid dynamics specifies airflows around flying aircraft. Other models are simulations that evaluate the interactions between entities over long periods of time. For example, astronomers have successfully simulated galactic collisions by treating galaxies as masses of particles representing stars. Because models make simplifications there is always a trade-off between model complexity and accuracy. The classical core abstraction of physics has been any natural process; in recent decades it expanded to include information processes and computers.

Mathematics.  Abstraction is the business of mathematics. Mathematicians are constantly seeking to identify concepts and structures that transcend physical objects. They seek to express the essential relationships among objects by eliminating irrelevant details. Mathematics is seen as supportive of all scientific fields. In 1931, Bertrand Russell wrote: “Ordinary language is totally unsuited for expressing what physics really asserts, since the words of everyday life are not sufficiently abstract. Only mathematics and mathematical logic can say as little as the physicist means to say.”

Anywhere you see a classification you are looking at a hierarchy of abstractions. Anywhere you see a theory you are looking at an explanation of how a set of abstractions interacts. Highly abstract concepts can be very difficult to learn because they depend on understanding much past history, represented in lower-level abstractions.

Differing Interpretations of the Same Abstractions

It is no surprise that different people have different interpretations about abstractions and thus get into arguments over them. After all, abstractions are mental constructs learned individually. Few abstractions have clear logical definitions as in mathematics or in object-oriented languages. Here are some additional examples showing how different fields approach differences of interpretation of their core abstractions.

Biology.  This is the science studying life. There is, however, no clear definition of life. How do biologists decide if some newly discovered organism is alive? They have agreed on a list of seven criteria for assessing whether an entity is living:

  • Responding to stimuli

  • Growing and developing

  • Reproducing

  • Metabolizing substances into energy

  • Maintaining a stable structure (homeostasis)

  • Structured from cells

  • Adaptability in changing environments

The more of these criteria hold for an organism, the more likely is a biologist to say that life is present.

Artificial intelligence.  Its central abstraction—intelligence—defies precise definition. Various authors have cited one of more of these indicators as signs of intelligence:

  • Passes IQ tests

  • Passes Turing test

  • Pinnacle of a hierarchy of abilities determined by psychologists

  • Speed of learning to adapt to new situations

  • Ability to set and pursue goals

  • Ability to solve problems

However, there is no agreement on whether these are sufficient to cover all situations of apparent intelligence. Julian Togelius has an excellent summary of the many notions of “intelligence” (and “artificial”) currently in play.6 This has not handicapped AI, which has produced a series of amazing technological advances.

Computer science.  Its central concept—information process—defies a precise definition. Among the indicators frequently mentioned are:

  • Dynamically evolving strings of symbols satisfying a grammar

  • Assessment that strings of symbols mean something

  • Mapping symbol patterns to meanings

  • Insights gained from data

  • Fundamental force in the universe

  • Process of encoding a description of an event or idea

  • Process of recovering encrypted data

  • Inverse log of the probability of an event (Shannon)

There is no consensus whether these are sufficient to cover all situations where information is present.

Neuroscience.  Consciousness is a core abstraction. Neuroscientists and medical professionals in general have agreed on a few, imprecise indicators of when someone is conscious.5 Some conscious people may fail all the indicators, and some unconscious people may satisfy some of the indicators. It may be impossible to ever know for sure if someone is conscious or not.

Business.  Innovation is a core abstraction. Business leaders want more innovation. Definitions vary from inventing new ideas, prototyping new ideas, transitioning prototypes into user communities, diffusing into user communities, and adopting new practice in user communities. Each definition is accompanied by a theory of how to generate more innovation. The definitions are sufficiently different that the theories conflict. There is considerable debate on which definition and its theory will lead to the most success.

Conclusion

The accompanying table summarizes the examples above. The “criteria” column indicates whether a field has a consensus on criteria for their core abstraction. The “explanatory” column indicates whether a field’s existing definitions adequately explain all the observable instances of their core abstraction. The “utility” column indicates whether they are concerned with finding applications of technologies enabled by their core abstraction.

Table. 

A few fields and their core abstractions
Field Abstraction Criteria? Explanatory? Utility?
Computing Information No No Yes
Physics Natural phenomena No Yes Yes
Mathematics Math concepts No Yes No
Biology Life Yes Yes Yes
Artificial Intelligence Intelligence No No Yes
Neuroscience Consciousness No Yes Maybe
Business Innovation No No Yes

Thus, it seems that the core abstractions of many fields are imprecise and, with only a few exceptions, the fields have no consensus on criteria to determine if an observation fits their abstraction. How do they manage a successful science without a clear definition of their core abstraction? The answer is that in practice they design systems and processes based on validated hypotheses. The varying interpretations are a problem only to the extent that disagreements generate misunderstanding and confusion.

A good way to bring out the differences of interpretation is to ask people how they assess if a phenomenon before them is an instance of their core abstraction. Thus you could say “Life is an assessment,” “intelligence is an assessment,” and so on. When you put it this way, you invite a conversation about the grounding that supports the assessment. For example, a biologist would ground an assessment that a new organism is alive by showing that enough of the seven criteria are satisfied. In other fields the request for assessment quickly brings out differences of interpretation. In business, for example, where there is no consensus on the indicators of innovation, a person’s assessments reveal which of the competing core abstractions they accept. That, in turn, opens the door for conversations about the value of each abstraction.

There is a big controversy over whether technology is dragging us into abstract worlds with fewer close relationships, fear of intimacy, and interaction limited to exchanges across computer screens. This is a particular problem for young people.3 Smartphones are intended to improve communication and yet users feel more isolated, unseen, unappreciated. Something is clearly missing in our understanding of communication, but we have not yet put our collective finger on it.

Two books may help sort this out. In Power and Influence, Nobel Prize economists Daron Acemoglu and Simon Johnson present a massive trove of data to claim that increasing automation often increases organizational productivity without increasing overall economic progress for everyone. They argue that the abstractions behind automation focus on displacing workers rather than augmenting workers by enabling them to take on more meaningful tasks.1 In How to Know a Person, David Brooks presents communication practices that help you see and appreciate the everyday concrete concerns of others.2

Maybe we need to occasionally descend from the high clouds of our abstractions to the concrete earthy concerns of everyday life.

]]>
https://cacm.acm.org/opinion/abstractions/feed/ 2 764610
The Sustainability Gap for Computing: Quo Vadis? https://cacm.acm.org/research/the-sustainability-gap-for-computing-quo-vadis/ https://cacm.acm.org/research/the-sustainability-gap-for-computing-quo-vadis/#respond Fri, 14 Feb 2025 20:23:27 +0000 https://cacm.acm.org/?post_type=digital-library&p=764872 A reformulated IPAT model provides insight for computer system engineers to consider computing's environmental impact.

]]>

Sustainability is undeniably a grand challenge. As the world population and the average affluence per person continues to grow, we are eagerly consuming Earth’s natural resources. The Earth Overshoot Day marks the date when the demand for ecological resources by humankind in a given year exceeds what the Earth can regenerate in a year’s time. While the world’s Earth Overshoot Day fell at the end of December in the early 1970s, it has progressively antedated since then and was computed to fall on Aug. 1 in 2024. The overshoot day is (much) earlier for many countries, as early as Feb. 26 for Singapore, March 13 for the U.S., March 26 for Canada, and April-May for most European countries as well as South Korea, Israel, Japan, and China.a

The continuously growing consumption of Earth’s resources, including materials and energy sources, (inevitably) induces climate change. Greenhouse gas (GHG) emissions result in detrimental global warming, and a recent study reports that the contribution of information and communication technology (ICT) to the world’s global GHG emissions, currently between 2.1% and 3.9%,11 is growing at rapid pace. While this percentage may seem small, it is not. In fact, ICT’s contribution to global warming is on par with (or even larger than) the aviation industry, which is estimated to be around 2.5%.b

Key Insights

  • Computing is responsible for a significant and growing fraction of the world’s global carbon footprint.

  • The status quo in which we keep per-device carbon footprint constant would lead to a 5.4× sustainability gap for computing relative to the Paris agreement within a decade

  • Meeting the Paris agreement for computing requires reducing the per-device carbon footprint by 15.5% per year under current population and affluence growth curves.

  • Based on a select number of published carbon footprint reports, it appears that while vendors indeed reduce the per-device carbon footprint, it does not seem to be enough to close the gap, urging our community to do more.

To combat global warming, the Paris Agreement under the auspices of the United Nations (UN) aims to limit global warming to well below 2 degrees Celsius, and preferably to 1.5 degrees Celsius, compared to pre-industrial levels. In 2019, the UN stated that  global emissions must be cut by 7.6% each year over the next decade to meet the Paris Agreement.c More recently in November 2023, the UN stated that insufficient progress has been made so far to combat climate change.d

Given the pressing need to act, along with the significant and growing contribution of computer systems to global warming, it is imperative that we, computer system engineers, ask ourselves what we can do to reduce computing’s environmental footprint within the socio-economic context. To do so, this article reformulates the well-known and widely used IPAT model,6 such that we can reason about the three contributing factors: population growth, increased affluence or number of computing devices per person, and carbon footprint per device over its entire lifetime, which includes the so-called embodied footprint for manufacturing, assembly, transportation, and end-of-life processing, and the operational footprint due to device usage during its lifetime.13

Growth in population and affluence leads to a growing sustainability gap.
Figure 1.  Growth in population and affluence leads to a growing sustainability gap. The widening sustainability gap for computing between the current status quo in per-device carbon footprint (which leads to a 2.4× increase in global carbon footprint within a decade) versus the Paris Agreement (which requires a 2.2× reduction). Closing the sustainability gap requires that we reduce the per-device carbon footprint by 15.5% per year under current population and affluence growth rates.

The growth in population and affluence leads to a growing sustainability gap as illustrated in Figure 1. If we were to keep the carbon footprint per device constant relative to the present time, the total carbon footprint due to ICT would still increase by 9.4% per year, leading to a 2.45× increase in GHG emissions over a decade. In contrast, meeting the Paris Agreement requires that we reduce GHG emissions by a factor 2.2×. Bridging this widening sustainability gap between the per-device status quo and the Paris Agreement requires that we reduce the carbon footprint per device by 15.5% per year or by a cumulative factor of 5.4× over a decade.

Analyzing the carbon footprint for a select number of computing devices (smartwatches, smartphones, laptops, desktops, and servers) reveals that vendors do pay attention to sustainability. Indeed, the carbon footprint per computing device tends to reduce in recent years, at least for some devices by some vendors. However, the reduction in per-device carbon footprint achieved in recent years appears to be insufficient to close the sustainability gap. The overall conclusion is that a concerted effort is needed to significantly reduce the demand for computing devices while reducing the carbon footprint per device at a sustained rate for the foreseeable future.

The IPAT Model

IPAT is the acronym of the well-known and widely used equation6 which quantifies the impact I of human activity on the environment:

I = P × A × T

P stands for population (that is, the number of people on earth), A accounts for the affluence per person or the average consumption per person, and T quantifies the impact of the technology on the environment per unit of consumption. The impact on the environment can be measured along a number of dimensions, including the natural resources and materials used (some of which may be critical and scarce); GHG emissions during the production, use, and transportation of products; pollution of ecosystems and its impact on biodiversity; and so on. The IPAT equation is used as a basis by the UN’s Intergovernmental Panel for Climate Change (IPCC) in their annual reports.

The IPAT equation has been criticized as being too simplistic, assuming that the different variables in the equation are independent of each other. Indeed, in contrast to what the above formula may suggest, improving one of the variables does not necessarily lead to a corresponding reduction in overall impact. For example, reducing T in the IPAT model by 50% through technology innovations to reduce the environmental impact per product does not necessarily reduce the overall environmental impact I by 50%. The fundamental reason is that a technological efficiency improvement may lead to an increase in demand and/or use, which in turn may lead to an increase, rather than a reduction, in overall impact. This is the well-known rebound effect or Jevons’ paradox, named after the English economist Williams Stanley Jevons, who was the first to describe this rebound effect. He observed that improving the coal efficiency of the steam engine led to an overall increase in coal consumption.2 Although there is no substantial carbon tax as of today, Jevons’ paradox still (indirectly) applies to computer systems. Efficiency gains increase a computing device’s compute capabilities, which stimulates its deployment (that is, more devices are deployed due to increase in demand) and its usage (that is, the device is used more intensively). The result may be a net increase in total carbon footprint across all devices despite the per-device efficiency gains.

The rebound effect can be (partly) accounted for in the IPAT model by expressing each of the variables as a compound annual growth rate (CAGR), defined as follows:

C A G R = V t V 0 1 / t 1

with V0 the variable’s value at year 0 and Vt its value at year t. The IPAT model can be expressed using CAGRs for the respective variables:

CAGR overall = i = 1 N ( C A G R i + 1 ) 1

This formulation allows for computing the annual growth rate in overall environmental impact or GHG emissions as a function of the growth rates of the individual contributing factors. If the growth rates incorporate the rebound effect, that is, higher consumption rate as a result of higher technological efficiency, the model can make an educated guess about the expected growth rate in environmental impact.3

The Environmental Impact of Computing

We now reformulate the IPAT equation such that it provides insight for computer system engineers to reason about the environmental impact of computing within its socio-economic context. We do so while focusing on GHG emissions encompassing the whole lifecycle of computing devices. The total GHG emissions C incurred by all computing devices on earth can be expressed as follows:

C = P × D P × C D

P is the world’s global population. D/P is a measure for affluence and quantifies the number of computing devices per capita on earth. C/D is a measure for technology and corresponds to the total carbon footprint per device. Note that C/D includes the whole lifecycle of a computing device, from raw material extraction to manufacturing, assembly, transportation, usage, and end-of-life processing. We now discuss how the different factors P, D/P, and C/D in the above equation scale over time.

Population.  The world population P has grown from one billion in 1800 to eight billion in 2022. The UN expects it to reach 9.7 billion in 2050 and possibly reach its peak at nearly 10.4 billion in the mid 2080s.e The world population annual growth rate was largest around 1963 with a CAGRP = +2.1%. Since then, the growth rate has reduced to around CAGRP = +0.9% according to the World Bank.f

Affluence.  The number of devices per person D/P increases at a fairly sharp rate7 (see Table 1). On average across the globe, the number of connected devices per capita increased from 2.4 in 2018 to 3.6 in 2023, or CAGRD/P = +8.4%. In the western world, that is, North America and Western Europe, the number of devices per person is not only a factor 2× to 4× larger than the world average, it also increases much faster with a CAGRD/P above +10%. The increase in the number of devices is in line with the annual increase in integrated circuits (ICs), that is, estimated CAGR = +10.2% according to the 2022 McClean report from IC Insights.18

Table 1.  Number of connected devices per capita.7
Region 2018 2023 CAGR
Global 2.4 3.6 +8.4%
Asia Pacific 2.1 3.1 +8.1%
Central and Eastern Europe 2.5 4.0 +9.9%
Latin America 2.2 3.1 +7.1%
Middle East and Africa 1.1 1.5 +6.4%
North America 8.2 13.4 +10.3%
Western Europe 5.6 9.4 +10.9%

Technology.  The carbon footprint per device C/D and its scaling trend CAGRC/D is much harder to quantify because of inherent data uncertainty and the myriad computing devices. The carbon footprint of a device depends on many factors, including the materials used, how those materials are extracted, how the various components of a device are manufactured and assembled, how energy efficient the device is, where these devices are used, the lifetime of the device, how much transportation is involved, how end-of-life processing is handled, and so on. Despite the large degree of uncertainty, it is instructive and useful to analyze lifecycle assessment (LCA) or product carbon footprint (PCF) reports that quantify the environmental footprint of a device. All LCA and PCF reports acknowledge the degree of data uncertainty; nevertheless, they provide invaluable information for consumers to assess the environmental footprint of devices.

To understand per-device carbon footprint scaling trends, we now consider a number of computing devices from different vendors. We leverage the carbon footprint numbers published in the products’ respective LCA or PCF reports. In particular, we use the resources available from Apple,g Google,h and Dell.i We now discuss carbon footprint scaling trends for smartwatches, smartphones, laptops, desktops, and servers.

The bottom line is that per-device carbon footprint has not increased dramatically over the past years and has even significantly decreased in some cases. Several interesting conclusions can be reached upon closer inspection across devices and vendors.

Smartwatches.  Figure 2 quantifies the carbon footprint for different generations of Apple Watches with similar capabilities (GPS versus GPS plus cellular) and sport band. All watches feature either an aluminum case (42mm in Series 1 to 3, 44mm in Series 4 to 6, and 45mm in Series 7 and 8) or a stainless case (Series 9). It is surprising perhaps to note that a smartwatch’s carbon footprint was on a rising trend until 2019 before declining. Indeed, the carbon footprint of a GPS watch has increased with a CAGRC/D = +23.9% from 2016 (Series 1) until 2019 (Series 5), while the carbon footprint of a GPS-plus-cellular watch has decreased with a CAGRC/D = −7.7% from 2019 (Series 5) until 2023 (Series 9).

Figure 2.  Carbon footprint for Apple Watches.

Smartphones.  Figure 3 illustrates the carbon footprint for Apple iPhones starting with iPhone 7 (release date in 2016) until iPhone 15 Pro Max (release date in 2023) with different SSD capacity. We note a similar trend for the Apple smartphones as for the smartwatches: Per-device carbon footprint increased until 2019, when it began declining. Indeed, from iPhone 8 (2017) to iPhone 11 Pro Max (2019) with 256GB SDD, the carbon footprint has increased from 71kg to 102kg CO2eq (CAGRC/D = +19.8%). From 2019 onward, we note a decrease in carbon footprint per device: From iPhone 11 Pro Max (2019) to iPhone 15 Pro Max (2023) with 512GB SDD, the carbon footprint decreased from 117kg to 87kg CO2eq (CAGRC/D = −7.1%). While Apple has been steadily decreasing the smartphone carbon footprint since 2019, it is worth noting that the declining trend is slowing down in recent years. For example, from iPhone 13 Pro Max (2021) to iPhone 15 Pro Max (2023) with 512GB SSD, the carbon footprint has decreased from 93kg to 87kg CO2eq (CAGRC/D = −3.3%).

Carbon footprint for Apple iPhones.
Figure 3.  Carbon footprint for Apple iPhones. Carbon footprint for Apple iPhones with different SSD capabilities (GB) from the iPhone 7 through the iPhone 15 Pro Max.

To analyze these trends across vendors, Figure 4 reports results for the Google Pixel phones; the plot reports carbon footprints for the nominal series (Pixel 2, 3, 4, 5, 6, 7, 8), the ‘a’ series (Pixel 3a, 4a, 5a, 7a, 8a), the XL series (Pixel 2XL, 3XL, 4XL), and the Pro series (Pixel 6Pro, 7Pro, 8Pro). As for Apple, we note a declining trend in recent years for Google smartphones: The per-device carbon footprint increased until mid 2021, after which it started trending downward. This is noticeable for the nominal series as well as for the high-end phone series (XL and Pro series). The decrease in carbon footprint since 2021 is substantial for the nominal series (CAGRC/D = −10.5%) and the Pro series (CAGRC/D = −8.8%), while remaining invariant for the ‘a’ series since mid 2021.

Figure 4.  Carbon footprint for Google Pixel smartphones.

Laptops.  Figure 5 reports the carbon footprint for Apple MacBook Pro and MacBook Air laptops with different configurations (screen size, see legend) as a function of their respective release dates; multiple laptops are reported per release date with different storage capacity, core count, and frequency. Several observations are worth noting. First, and perhaps not surprisingly, MacBook Air laptops incur a smaller carbon footprint than the more powerful MacBook Pro laptops. Second, for a given screen size, we note a steady decrease in carbon footprint, for example, MacBook Pro 16-in. (CAGRC/D = −6.9% from 2019 to 2023) and MacBook Air 13-in. (CAGRC/D = −5.1% from 2018 to 2024). Third, while this continuous decrease in per-device carbon footprint is encouraging, there is a caveat: Discontinuing a particular laptop configuration and replacing it with a more powerful device comes with a substantial carbon footprint increase. In particular, replacing the MacBook Pro 15-in. with a 16-in. configuration mid-2019 increases the carbon footprint by at least 11.3%; likewise, the transition from 13-in. to 14-in. in the second half of 2022 led to an increase of at least 33.5% for entry-level MacBook Pro laptops.

Figure 5.  Carbon footprint for Apple MacBook Pro and Air laptops with different screen configurations.

Looking at Dell, we note a slightly different outcome. The data in Figure 6 reports the carbon footprint for the 3000, 5000, and 7000 Dell Precision laptops. The per-device carbon footprint increased from 2018 until 2023 for the 5000 (CAGRC/D = +4.1%) and 7000 (CAGRC/D = +3.8%) laptops, while being invariant for the 3000 laptops. Note that the carbon footprint drastically drops for the most recent laptops released in February and March 2024, but this is due to a change in carbon accounting from MIT’s PAIA tool to Dell’s own ISO14040-certified LCA tool.

Figure 6.  Carbon footprint for Dell Precision laptops.

Desktops and workstations.  The outlook is mixed for desktops and workstations, with some trends increasing and others decreasing. Figure 7 shows carbon footprint for the Dell OptiPlex 700 Series Tower desktop machines (left) decreasing at a rate of CAGRC/D = −8.1% but increasing at a rate of CAGRC/D = +4.0% for Dell Workstations 5000 and 7000 Series (right).

Figure 7.  Carbon footprint for Dell desktops and workstations.

Servers.  Figure 8 reports the carbon footprint for Dell PowerEdge rackmount ‘R’ servers across four generations (13th, 14th, 15th, and 16th); this includes Intel- and AMD-based systems. Server carbon footprint numbers are subject to its specific configuration and deployment, more so than (handheld) consumer devices for at least two reasons. First, because the operational footprint tends to dominate for servers—unlike handheld devices which are mostly dominated by their embodied footprint13—the location of use (and its power-grid mix) has a substantial impact. Second, hard-drive capacity, memory capacity, and processor configuration heavily impact the overall carbon footprint. Overall, server carbon footprint seems to be relatively constant over the past decade, although we note a small increase in average carbon footprint from the 13th to the 16th generation (CAGRC/D = +1.8%). The carbon footprint of a typical high-end server tends to range between 8,000kg and 15,000kg CO2eq over the past decade. Entry-level servers tend to have a lower carbon footprint below 6,000kg CO2eq with a downward trend in recent years. (See the couple of data points in the bottom right corner in Figure 8.)

Figure 8.  Carbon footprint for Dell PowerEdge rackmount ‘R’ servers.

Discussion.  It is (very) impressive to note that the compute power of computing devices has dramatically increased over the past years while not dramatically increasing the per-device carbon footprint. In fact, for several computing devices, we note a decreasing trend in per-device carbon footprint—see also Table 2 for a summary—especially in recent years, which is particularly encouraging to note.

Table 2.  Per-device carbon footprint scaling trends.
Device Model Period CAGR
Smartwatch Apple Watch 2019–2023 -7.7%
Smartphone Apple iPhone Pro Max 2019–2023 -7.1%
Apple iPhone Pro Max 2021–2023 -3.3%
Google Pixel 2021–2023 -10.5%
Laptop Apple MacBook Pro 16-in. 2019–2023 -6.9%
Apple MacBook Air 13-in. 2018–2024 -5.1%
Dell Precision 7000 2018–2023 +3.8%
Desktop Dell OptiPlex 700 2019–2022 -8.1%
Dell Workstations 5000 and 7000 2018–2023 +4.0%
Server Dell PowerEdge rackmount 2014–2024 +1.8%

One may wonder whether the recent reduction in per-device carbon footprint comes from reductions in embodied or operational footprint. Upon closer inspection, it turns out that the key contributor to the total reduction in carbon footprint varies across device types. For the Apple smartwatches, the relative decrease in embodied footprint (−7.4% per year) is more significant than the decrease in operational footprint (−2.8% per year). Also, for the MacBook Pro 16-in. laptops, the embodied footprint has decreased at a faster pace (−7.9% per year) than the operational footprint (−3.5% per year). In contrast, for the iPhone Pro Max smartphones, we note a more significant reduction in operational footprint (−11.2% per year) than in embodied footprint (−5.7% per year). For the MacBook Air 13-in. laptops, we even note an increase in operational footprint (+14.9%) while the embodied footprint trends downward (−5.6%).

Overall, per-device carbon footprint decreases for most of the devices analyzed in this work, and in cases where it increases, the increase is limited. The reason for these trends is mixed. The question now is whether this overall declining trend in per-device carbon footprint is sufficient to reduce the overall environmental footprint of computing, and, even better, for meeting the Paris Agreement, which we discuss next.

Quantifying the Sustainability Gap

Recall that population and affluence increase, (CAGRP = +0.9%) and (CAGRD/P = +8.4%), respectively. Technology, on the other hand, seems to decrease for many devices, ranging from CAGRC/D = −3.3% to −10.5%, while increasing for others from +1.8% to +4.0%, as summarized in Table 2. The question now is whether these trends lead to an overall increase or decrease in the environmental footprint of computing.

Figure 9 predicts the overall carbon footprint for the next decade normalized to present time for a variety of typical per-device carbon footprint scaling trends, that is, CAGRC/D = +4%, −5%, −10%. In addition, we consider the following three scenarios:

Total carbon footprint normalized to present time.
Figure 9.  Total carbon footprint normalized to present time. Total carbon footprint normalized to present time for different per-device carbon footprint scaling trends and scenarios (see CAGRC/D values in the legend).

Scenario #1: Status quo per-device footprint.  If we were to keep the carbon footprint per device constant relative to present time, that is, CAGRC/D = 0%, the total carbon footprint would still increase substantially (CAGRC = +9.4%). This is simply a consequence of the growing population and the increasing affluence or number of computing devices per person. Because this is an exponential growth curve, this implies that the total carbon footprint of computing would increase by a factor 2.45× over a decade. In other words, even if we were to keep the carbon footprint per device constant, the total carbon footprint of computing would still dramatically increase.

Scenario #2: Status quo overall footprint.  If we want to keep the overall carbon footprint of computing constant relative to the present time, that is, CAGRC = 0%, we need to reduce the carbon footprint per device by CAGRC/D = −8.6% per year. This is to counter the increase in population and number of devices per person. Reducing the carbon footprint per device by 8.6% year after year for a full decade is a non-trivial endeavor. To illustrate how challenging this is, consider a device that incurs a carbon footprint of 100kg CO2eq. Reducing by 8.6% per year requires that the carbon footprint is reduced to 40.6kg CO2eq within a decade; in other words, the carbon footprint needs to reduce by more than a factor 2.4× over a period of 10 years.

Scenario #3: Meeting the Paris Agreement.  To make things even more challenging, to meet the Paris Agreement, we need to reduce global GHG emissions by a factor 2.2× over a decade or by 7.6% per year, that is, CAGRC = −7.6%. To achieve this, we would need to reduce the carbon footprint per device by 15.5% per year, that is, CAGRC/D = −15.5%. This implies that we need to reduce the carbon footprint per device by a factor 5.4× over a decade.

It is clear from Figure 9 that the impact of the per-device carbon footprint scaling trend has a major impact on the overall environmental footprint. Relatively small differences in CAGR lead to substantial cumulative effects over time due to the exponential growth curves. In particular, the status quo per-device carbon footprint (CAGRC/D = 0%) leads to a 2.45× increase in overall carbon footprint over a decade, while the Paris Agreement requires that we reduce the total carbon footprint by 2.2× (CAGRC/D = −15.5%). Even if we were to reduce the per-device carbon footprint at a relatively high rate (CAGRC/D = −8.6%) to maintain a status quo in total carbon footprint, the gap with the Paris Agreement would still increase at a rapid pace.

As noted from Table 2, most devices do not follow a trend that complies with these required trends: Reported per-device carbon footprint CAGRs are not anywhere close to the required CAGRC/D = −15.5% (to meet the Paris Agreement) nor do they uniformly meet the CAGRC/D = −8.6% (to keep total carbon footprint constant relative to present time). To close the sustainability gap, one needs to reduce the per-device carbon footprint by 15.5% per year for the next decade. This implies that the computing industry should do more to keep its carbon footprint under control. This leads to the overall conclusion that there is a substantial gap between the current state of affairs versus meeting the Paris Agreement. Bridging the sustainability gap is a non-trivial and challenging endeavor, which will require significant innovation in how we design and deploy computing devices beyond current practice.

The Socio-Economical Context

The above analysis assumes that the world population and the number of devices per person will continue to grow at current pace for the foreseeable future. The task of decreasing the carbon footprint per device by 15.5% per year to meet the Paris agreement can be loosened to some extent by embracing a certain level of sobriety in affluence, that is, limiting the number of devices per person. This requires a perspective on the socio-economical context of computing, which includes economic business models, regulation, and legislation.

The computing industry today is mostly a linear economy where devices are manufactured, used for a while, and then discarded. The lifetime of a computing device can be relatively short, for example, two to four or five years, leading to increased e-waste. Reusing, repairing, refurbishing, repurposing, and remanufacturing devices could contribute to a circular economy in which the lifetime of a computing device is prolonged, thereby reducing e-waste and tempering the demand for more devices.10 For example, Switzer et al.23 repurpose discarded smartphones in cloudlets to run microservice-based applications. Reducing the demand for devices could possibly relax the need for reducing per-device carbon footprints.

There is a moral aspect associated with reducing the demand for devices, which is worth highlighting. As mentioned previously, affluence is higher in the western world (North America and Western Europe) compared to other parts of the world and moreover it is increasing faster. From an ethical perspective, this suggests that the western world should make an even greater effort to reduce the environmental footprint of computing—in other words, we should not necessarily expect other parts of the world to make an equally big effort to solve a problem the western world is mostly responsible for. This implies that the western world should step up its effort in embracing sobriety (that is, consume fewer devices per person) and making individual computing devices even more sustainable.

In addition to transitioning toward a circular economy, other business models can also be embraced. Today, most cloud services are free to use, see for example social media, mail, Web search, and so on, while relying on massive data collection. Maintaining, storing, processing, and searching Internet-scale datasets requires massive compute, memory, and storage resources. According to a recent study by the International Energy Agency,15 datacenters are estimated to account for about 2% of the global electricity usage in 2022; and by 2026, datacenters are expected to consume 6% of the nation’s electricity usage in the U.S. and 32% in Ireland. Moreover, data storage incurs a substantial embodied footprint.24 The environmental footprint of free Internet services is hence substantial. Allowing low-priority files to degrade in quality over time could possibly temper the environmental cost for storage devices.26 But we could go even further by changing the business and usage models of Internet-scale services. Imposing a time restriction for uploaded content could possibly temper the demand for more processing power and storage capacity. We may want to limit how long we keep data around depending on its usefulness and criticality. To make it concrete: Do we really need to keep (silly) cat videos on the Web forever? Limiting to a day or a week may serve the need. Alternatively, or complementarily, one could demand a financial compensation from the customer for using online services. In particular, one could ask customers if they are willing to pay to keep their content online. For example, do you want to pay for your cat videos to remain online for the next month or year?

Transitioning to renewable energy sources (solar, wind, hydropower) is an effective method to reduce the carbon footprint of computing—as with any other industry. Renewables during chip manufacturing have the potential to drastically reduce a device’s embodied carbon footprint. Conversely, renewables at the location of device use drastically reduce a device’s operational emissions. This is happening today as renewables take up an increasingly larger share in the electricity mix.20 However, there are several caveats. First, total electricity demand increases faster than what renewables can generate, increasing the reliance on brown electricity sources (that is, coal and gas) in absolute terms.20 In other words, the transition rate to renewables is not fast enough to compensate for the increase in population and affluence. Second, the amount of green energy is too limited to satisfy all stakeholders. For example, Ireland has decided to limit datacenter construction until 2028 because allowing more datacenters to be deployed would compromise the country’s commitment that 80% of the nation’s electricity grid should come from renewables by 2030.16 Third, while renewables reduce total carbon footprint, it does not affect other environmental concerns, including raw material extraction, chemical and gas emissions during chip manufacturing, water consumption, impact on biodiversity, and so on.

The analysis performed in this paper considered computing as a standalone industry. But computing may enable other industries to become more sustainable, thereby (partially) offsetting its own footprint. This could potentially lead to an overall reduction in environmental footprint.19 For example, computer vision enables more efficient agriculture using less water resources and pesticides; artificial intelligence and machine learning could make transportation more environmentally friendly; smart grids that use digital technologies could increase the portion of renewables in the electricity mix in real time; or Internet-of-Things (IoT) devices could help reduce emissions in residential housing. While the anticipated sustainability gains in other industries may be substantial, one must be careful when analyzing such reports, that is, one has to carefully understand the assumptions and the associated limitations to fully grasp the validity of such analyses.21 Moreover, one must be wary of Jevons’ paradox as mentioned before: Making a product or service more carbon-friendly may lead to an increase in overall carbon footprint if the efficiency gain leads to increased deployment and/or usage. In other words, one should be aware of the bigger picture—unfortunately, holistic big-picture assessments are extremely complicated to make and anticipate.

Finally, regulation and legislation may be needed to temper the environmental footprint. The previously cited IEA report15 states that “regulation will be crucial in restraining data center energy consumption” while referring to the European Commission’s revised energy-efficiency directive. The latter entails that datacenter operators have to report datacenter energy usage and carbon emissions as of 2024, and they have to be climate-neutral by 2030. Further, the European Parliament adopted the so-called ‘right to repair’ directive, which requires manufacturers to repair goods with the goal of extending a product’s lifetime—thereby reducing e-waste and the continuous demand for new devices. Overall, innovation in regulation, legislation, and/or business models will be needed to incentivize (or even force) manufacturers, operators, and customers to temper the demand for more devices, while making sure that our computing industry can still thrive and generate welfare. This is a call for action for our community to reach out to psychologists, sociologists, law and policy makers, entrepreneurs, business people, and so on to holistically tackle the growing environmental footprint of computing.

Related Work

Our computer systems community recently started considering sustainability as a design goal, and prior work focused mostly on characterizing,8,12,13,24 quantifying,9,14,17 and reducing1,4,5,22,25 the carbon footprint per device. However, as argued in this paper, to comprehensively and fully understand and temper the environmental footprint of computing, one needs to consider the socio-economic context within which we operate. Population growth and increased affluence is a current reality we should not be blind to and which impacts what we must do to reduce the overall environmental impact of computing.

Conclusion

This paper described the sustainability gap and how it is impacted by population growth, the increase in affluence (increasing number of devices per person), and the carbon intensity of computing devices. Considering current population and affluence growth, the carbon intensity of computing devices needs to reduce by 9.4% per year to keep the total carbon footprint of computing constant relative to present time, and by 15.5% per year to meet the Paris Agreement. Several case studies illustrate that while (some) vendors successfully reduce the carbon footprint of devices, it appears that more needs to be done. A concerted effort in which both the demand for electronic devices and the carbon footprint per device is significantly reduced on a continuous basis for the foreseeable future, appears to be inevitable to keep the rising carbon footprint of computing under control and, if possible, drastically reduce it.

]]>
https://cacm.acm.org/research/the-sustainability-gap-for-computing-quo-vadis/feed/ 0 764872
The Impact of AI on HRMS https://cacm.acm.org/blogcacm/the-impact-of-ai-on-hrms/ https://cacm.acm.org/blogcacm/the-impact-of-ai-on-hrms/#respond Wed, 12 Feb 2025 19:54:16 +0000 https://cacm.acm.org/?p=764718 HR has shifted from performing administrative tasks to using analytics to develop more productive workers and leaders.

]]>
The pace of adoption of new technologies in the world of Human Resources (HR) is moving at an astounding speed. Automation and artificial intelligence (AI) within human resource software have taken over admin work and tedious tasks.

Besides these applications, advances in machine learning, chatbots, computer vision, and other emerging AI solutions are increasingly replacing human decisions at scale.

What are the implications of integrating technology into the core functions of HR, and how will these changes reshape roles and operations? 

1) How Is AI Changing HR?

Emerging technologies have had a transformative impact on all aspects of businesses, including the human resource department. Companies aim to provide faster services, improve quality, and increase savings.

It’s changed the HR function and how it operates.

Today, we use AI to support a culture of collecting and analyzing data to make better-informed decisions and use these insights for continuous development.

The focal point of HR has shifted from admin tasks to utilizing analytics to create a more productive workforce and develop future leaders.

a. Automation

The most straightforward answer to how AI is changing Human Resources Management Systems (HRMS) is the automation of standardized processes. According to Deloitte, as much as 57% of HR’s time is spent on administrative tasks, instead of strategic planning.

Automation enables HR professionals to have more meaningful interactions, reduce errors, and ease the administrative workload. The latest AI-powered tools are designed to streamline workflow processes, seamlessly integrate them, and support all areas in HR’s scope. 

b. Reducing Bias

Another growing application of AI is the promise to reduce unconscious bias by disregarding information such as applicants’ age, race, and gender. This cognitive selection bias can seep into many areas like recruitment, decision-making, or performance evaluations. 

When organizations make data-driven decisions, they send a message of integrity. However, as these algorithms learn by following previous patterns of behavior by recruiters, they should not be left unchecked. AI offers a way to keep decision-making objective by giving candidate recommendations based on performance and knowledge. 

c. Retention

Employers can gain deep insights into their workforce and introduce various data-driven HR initiatives to improve retention by using cutting-edge technology.

AI solutions can be applied in every phase of the entire employee lifecycle. From personalized learning and development programs, using wearables to monitor wellness, to gamifying work to check employee engagement. 

d. Talent Acquisition

Time is of the essence when recruiting and hiring new employees. Recruiters can optimize talent acquisition by using AI to eliminate time-consuming and repetitive tasks. For example, implementing applicant tracking systems saves valuable time and helps hire more efficiently, plus finds and screens candidates faster and with more fairness.

2) Data Security in HRMS

Any major operational change will bring forth some disruption and inevitable risks, and the digital transformation in HR is no different. 

With the move to the cloud and centralization of employee data storage within AI-powered HRMS solutions, companies must ensure the confidentiality and integrity of their data.

Storing sensitive employee information online like banking details, Social Security numbers, and addresses heightens a company’s risk related to data breaches and cybercrime, and possibly substantial losses. Statistics show the cost of a data breach to an American company averaged $9.36 million in 2024. 

However, many businesses are using security AI and automation to detect and respond to these attacks, whether through encryption, anomaly detection, or multi-factor authentications. In fact, the average cost savings for organizations that used AI to protect their data versus those that didn’t is around $2.2 million

3) Integrating AI with HRMS

The modern workforce requires data science to address some of the most urgent challenges and provide innovative solutions to companies. After merging AI with HRMS systems, employers can experience an all-encompassing system that assists them in various internal HR functions.

Candidate engagement, training, onboarding, recruitment, talent management, and ERM (Employee Relationship Management) are all optimized after implementing AI in the HRMS work system.

a. Speedy Application Screening

Recruiters can now engage candidates before or after applying for a position in an organization. With the help of AI and chatbots, companies can test candidates, answer common questions about the role, address specific concerns, and get feedback and information about the candidate. Natural language processing is accelerating and becoming more sophisticated, providing candidates with a personalized, human-like experience. 

b. Candidate Engagement Process

As a large part of the recruitment process is automated, the focus of HR teams shifts from operational tasks to improving the candidate experience, creating engagement, and attracting candidates to apply. 

Artificial intelligence streamlines the candidate engagement process beyond the standard automated emails and messaging workflows. This engagement can be real-time and personalized to the individual candidate with AI.

In addition, AI can help give candidates a quick insight into a company’s work culture and values. 

c. Re-engagement of Candidates

AI allows you to re-engage a group of candidates from your database and gauge their interest in a new position or role. Most AI platforms allow recruiters to nurture candidate relationships. This means that candidate records can be stored in a database for future use—this way, the candidate does not need to resubmit their application when applying for several positions.

d. Fast Onboarding

The first day of a job and new-hire orientation can be overwhelming for many new employees. Incorporating artificial intelligence into the onboarding process can help employees learn about company policies, procedures, and benefits coverage, while verifying employment documentation.

e. Effective Workforce Development

Adopting AI solutions in the HRMS system can significantly improve the employee experience by identifying which employees need training or new skill sets. AI offers employees a personalized approach to their career development program or even coaching guided by coach bots for every individual. 

It’s not surprising then that 96% of large and mid-size companies and 81% of small companies already use a learning management system.

VR and AR technology will also play a valuable role in the learning and development of workers, especially in high-risk roles, where skills can be developed in safer environments. 

f. Improved ERM

An ERM module in HRMS systems provides an intelligent solution to track employees, their work relations, and all reporting and compliance processes. HR teams can structure the workforce into organizational units, define the reporting relations between managers and employees, and align payroll. 

4) The Future of HRMS

HRMS tools must consistently upgrade and integrate AI to create better workplaces to match this rapidly changing world. The transformation AI can bring to organizations is seen across various processes, and AI’s role in the workplace will only keep increasing. 

The pandemic showed that AI-driven HRMS can quickly adapt to flexible and hybrid work environments, with innovative tools being used for collaboration or intelligent scheduling.

Today, many organizations have adopted cloud-based and mobile-compatible HR management systems, so employees can easily help themselves with self-service anywhere and on any device. 

Even beyond that, AI-powered tools are transforming physical environments with desk and meeting room booking systems in an effort to improve the employee experience.

Office layouts, occupancy sensors, and optimizing the temperature and air quality are just some of the real-life applications to better utilize the workplace and make it more comfortable.

5) Conclusion

Apart from simplifying HR’s role, AI helps steer the direction of business strategies, enforces more informed decision-making, and helps overcome significant disruptive challenges. Businesses that see the value of introducing AI to their systems in a highly competitive market will win the war for talent, improve efficiency, and gain a significant advantage.

Sources

Modernizing HR: Design Thinking and New Technologies to Help Enhance Employee Experience (Deloitte)

Average cost of a data breach in the United States from 2006 to 2024 (Statista)

Cost of a Data Breach Report 2024 (IBM)

2023 Training Industry Report (Training Magazine)

Alex Tray is a system administrator and cybersecurity consultant with 10 years of experience. He is currently self-employed as a cybersecurity consultant and as a freelance writer.

]]>
https://cacm.acm.org/blogcacm/the-impact-of-ai-on-hrms/feed/ 0 764718
California’s AI Act Vetoed https://cacm.acm.org/opinion/californias-ai-act-vetoed/ https://cacm.acm.org/opinion/californias-ai-act-vetoed/#respond Mon, 03 Feb 2025 22:18:50 +0000 https://cacm.acm.org/?post_type=digital-library&p=764612 Should governments decide what regulations are necessary to ensure safe development and deployment of AI technologies?

]]>

Concerns that artificial intelligence (AI) systems pose serious risks for public safety have caused legislators and other policymakers around the world to propose legislation and other policy initiatives to address those risks. One bold initiative in this vein was the California legislature’s enactment of SB 1047, the Safe and Secure Innovation for Frontier Artificial Intelligence Models Act, in late August 2024.

Lobbying for and against SB 1047 was so intense that California’s governor Gavin Newsom observed that the bill “had created its own weather system.” In the end, the governor vetoed the bill for reasons I explain here. After a brief review of the main features of SB 1047, this column points out key differences between SB 1047 and the EU’s AI Act, identifies key supporters and opponents of SB 1047, and discusses arguments for and against this bill. It also explains Governor Newsom’s reasons for vetoing that legislation and considers whether national or state governments should decide what AI regulations are necessary to ensure safe development and deployment of AI technologies.

Key Features of SB 1047

Under SB 1047, developers of very large frontier models (defined as models trained on computing power greater than 1026 integer or floating point operations or costing more than $100 million at the start of training) and those who fine-tune large frontier models (also measured by compute requirements and/or training costs) would be responsible to ensure that these models will not cause “critical harms.”

The bill identifies four categories of critical harms:

  • Creation or use of chemical, biological, radiological, or nuclear weapons causing mass casualties;

  • Mass casualties or more than $500 million damage because of cyberattacks on critical infrastructure;

  • Mass casualties or more than $500 million in damages resulting in bodily injury or damage to property that would be a crime if humans did it; and

  • Other comparably grave harms to public safety and security.

Under this bill, developers of large frontier models would be required to take numerous steps at three phases of development: some before training, some before use of such a model or making it available, and some during uses of covered models. Among the required steps would be installing a “kill switch” at the pre-training stage, taking reasonable measures to prevent models from posing unreasonable risks, and publishing redacted copies of the developers’ safety and security protocols. (A “kill switch” would enable humans to stop an AI system from becoming an autonomous actor capable of inflicting critical harms.)

Developers would also be required to hire independent third-party auditors to ensure compliance with the law’s requirements. They would further be obliged to submit these audits and a statement of compliance annually with a state agency. Developers would further be responsible for reporting any safety incident of which they become aware to that agency within 72 hours of learning about it.

The legislation authorized the California Attorney General to file lawsuits against frontier model developers who violated that law’s requirements seeking penalties for up to 10% of the initial cost of model development for a first violation and up to 30% of development costs for subsequent violations. Whistleblowers who called attention to unreasonable risks that frontier models pose for causing critical harms would be protected against retaliation.

In addition, SB 1047 would authorize establishment of a new California agency to publish implementing guidelines for compliance with the Act. This agency would have received the required audit and compliance reports, overseen model development, and proposed amendments as needed (including updates to the compute thresholds).

Comparing SB 1047 to EU’S AI Act

SB 1047 and the European Union’s AI Act both focus on safety issues posed by advanced AI systems and risks that AI systems could cause substantial societal harms. Both require the development of safety protocols, pre-deployment testing to ensure systems are safe and secure, and reporting requirements, including auditing by independent third parties and compliance reports. Both would impose substantial fines for developers’ failure to comply with the acts’ safety requirements.

There are, however, significant differences between SB 1047 and the EU AI Act. For one thing, SB 1047 focused its safety requirements mainly on the developers of large frontier models rather than on deployers. Second, the California bill focused secondarily on those who fine-tune large frontier models, not just on initial developers. The AI Act does not address fine-tuning.

Third, SB 1047 would require developers to install a “kill switch” so that the models can be turned off if the risks of critical harms are too great. The EU’s AI Act does not require this. Fourth, the California bill assumed that the largest models are those that pose the most risks for society, whereas the AI Act does not focus on model size. Fifth, SB 1047 was intended to guard against those four specific types of critical harms, whereas the EU’s AI Act has a broader conception of harms and risks that AI developers and deployers should design to avoid.

Proponents of SB 1047

Anthropic was the most prominent of the AI model developers to have endorsed SB 1047. (Its support came after the bill was amended to drop a criminal penalty provision and to substitute a “reasonable care” instead of a “reasonable reassurance” standard for the duty of care expected of large frontier model developers). Thirty-seven employees of leading AI developers expressed support for SB 1047 as well.

Yoshua Bengio, Geoff Hinton, Stuart Russell, Bin Yu, and Larry Lessig are among the prominent proponents of SB 1047 as a “bare minimum effective regulation.” They believe that making developers of advanced frontier models responsible for averting critical harms is sound because these developers are in the best position to prevent such harms.

Proponents consider SB 1047 to be “light touch” regulation because it does not try to control design decisions or impose specific protocols on developers. They believe that the public will not be adequately protected if malicious actors are the only persons or entities that society can hold responsible for grave harms.

The AI Policy Institute reported that 65% of Californians support SB 1047 and more than 80% agree that advanced AI system developers should have to embed safety measures in the systems and should be accountable for catastrophic harms. Proponents further believe that SB 1047 will spur significant research and advance the state of the art in safety and security of AI models.

Without this new regulatory regime, moreover, proponents believe developers who are willing to invest in safety and security will be at a competitive disadvantage to firms that cut corners on safety and security design and testing to get to market faster.

Opponents of SB 1047

Google, Meta, and OpenAI, along with associations of technology companies, as well as Marc Andreesen and Ben Horowitz, opposed SB 1047 in part because it focused on the development of models instead of on harmful uses of such models. These opponents are concerned this law will impede innovation and American competitiveness in AI industries.

OpenAI argued that because SB 1047 heavily emphasizes national security harms and risks, it should be for the U.S. Congress, not the California legislature, to regulate AI systems to address these kinds of harms.

Among SB 1047’s opponents are many AI researchers, including notably Professors Fei Fei Li of Stanford and Jennifer Chayes of UC Berkeley. These researchers are concerned about the bill’s impacts on the availability of advanced open models and weights to which researchers want access and on which they want to build.

San Francisco Mayor London Breed and Congresswomen Nancy Pelosi and Zoe Lofgren were among the other prominent critics of SB 1047. Lofgren, who serves on a House subcommittee focused on science and technology issues, wrote an especially powerful letter to Governor Newsom expressing her reasons for opposing that bill. Among other things, Lofgren said that AI regulations should be based on demonstrated harms (such as deep fakes, misinformation, and discrimination), not hypothetical ones (such as those for which kill switches might be needed).

The science of AI safety, noted Lofgren, is very early stages. The technical requirements that SB 1047 would impose on developers of large frontier models are thus premature. While the National Institute of Science and Technology aims to develop needed safety protocols and testing procedures, these measures are not yet in place. Nor are voluntary industry guides yet fully developed.

Lofgren also questioned SB 1047’s “kill switch” requirement. Although this might sound reasonable in theory, such a requirement would undermine the development of ecosystems around open models. She agreed with a report of the National Telecommunications and Information Administration that there is insufficient evidence of heightened risks from open models to justify banning them.

Lofgren also expressed concern about innovation arbitrage. If California regulates AI industries too heavily or in inappropriate ways, it might lose its early leadership in this nascent industry sector. And U.S. competitiveness would be undermined.

Governor Newsom’s Reactions

Governor Gavin Newsom issued a statement explaining his reasons for vetoing SB 1047. He pointed out that California is home to 32 of the world’s leading AI companies. He worried that this law would harm innovation in California’s AI industries. Regulation should, he believes, be based on empirical evidence and science.

Newsom questioned whether the cost and amount of computing power needed for AI model training is the right regulatory threshold. He suggested it might be better to evaluate risks based on ecosystems in which AI systems were deployed or on uses of sensitive data. He warned that the bill’s focus on very large models could give the public a false sense of security because smaller models may be equally or more dangerous as the ones SB 1047 would regulate. While recognizing the need for AI regulations to protect the public, Newsom observed that the AI technology industry is still in early stages and regulations need to be balanced and able to be adapted as the industry matures.

The governor agreed with SB 1047’s sponsors that it would be unwise to wait for a catastrophe to protect the public from AI risks and that AI firms should be held accountable for harms to which they have contributed. But SB 1047, in his view, was just not the right law at the right time for California.

To demonstrate his commitment to ensuring proper attention to public safety, Governor Newsom appointed an expert committee of thought leaders to advise him further about how California can achieve the delicate policy balance between promoting the growth of AI industries and research communities and protecting the public against unreasonable risks of harm. Joining Fei Fei Li and Jennifer Chayes on this committee is Tino Cuellar, a former Stanford Law professor, a former California Supreme Court Justice, and now Executive Director of the Carnegie Institute for Peace.

Despite vetoing SB 1047, the governor signed into law 19 other AI-related bills passed by the California legislature this year. Two of them regulate deep fakes, one obliges developers to make disclosures about AI training data, and one requires provenance data for AI-generated outputs.

Conclusion

The sponsors of SB 1047 seem to have carefully listened to and heeded warnings of some prominent computer scientists who are deeply and sincerely worried about AI systems causing critically serious harms to humankind. However, there is no consensus among scientists about AI public safety risks.

Concerns that advanced AI systems, such as HAL in 2001: A Space Odyssey, will take over and humans will not be able to stop them because their developers failed to install kill switches seem implausible. Legislation to regulate AI technologies should be based on empirical evidence of actual or imminent harms, not conjecture.

In any event, regulation of AI systems that pose risks of national security harms would optimally be done at the national, not state, level. But the Trump Administration is less likely than the Biden Administration to focus on systemic risks of AI, so maybe the state of California should lead the way in formulating a regulatory regime to address these risks.

]]>
https://cacm.acm.org/opinion/californias-ai-act-vetoed/feed/ 0 764612
Staffing for Semiconductors https://cacm.acm.org/news/staffing-for-semiconductors/ https://cacm.acm.org/news/staffing-for-semiconductors/#respond Thu, 30 Jan 2025 18:22:03 +0000 https://cacm.acm.org/?p=764517 An expected shortfall in skilled workers poses a risk to the U.S.'s ability to compete in the global semiconductor manufacturing and chip design industry, the SIA says.

]]>
With advanced technologies such as artificial intelligence (AI), 5G telecommunication, and the Internet of Things (IoT) increasing the demand for semiconductors, and to ensure the success of The CHIPS and Science Act and drive innovation in the sector, the U.S. needs a skilled workforce.

To help secure the semiconductor industry in the U.S., the Biden Administration in November finalized a $6.6-billion award to Taiwan Semiconductor Manufacturing Company (TSMC) to construct chip factories in Phoenix, AZ. The world’s largest chipmaker and supplier to major technology companies, TSMC plans to invest $65 billion to build three semiconductor fabrication plants in the state. The facilities are expected to create thousands of jobs and strengthen the country’s position in the global semiconductor supply chain.

Revenue in the U.S. semiconductor market is expected to show an annual growth rate of 9.40% between 2024 and 2029, resulting in a market volume of $131 billion by 2029, according to global data and business intelligence platform Statista. Now the question is, will there be enough skilled workers to fill these positions to make the domestic semiconductor industry successful?

A 2024 report from the Semiconductor Industry Association (SIA) found that the semiconductor industry workforce will grow by almost 115,000 jobs by 2030 (from approximately 345,000 jobs now to approximately 460,000 jobs by the end of the decade). Of those new jobs, 67,000 or 58% risk going unfulfilled at current degree completion rates.

“The U.S. is expected to fall significantly short of the demand for skilled workers,” the SIA report warned. This includes 13,400 computer scientists and 26,400 technicians. Failure to address this gap in skilled workers poses risks to the country’s ability to compete in the global economy in semiconductor manufacturing and chip design, the SIA said.

More than one million additional skilled workers will be needed by 2030, according to Deloitte. However, about 60% of new manufacturing jobs in the U.S. semiconductor industry will not require a four-year college degree, the SIA report noted.

The Deloitte report suggests talent could be sourced from friendshoring regions, recruiting skilled immigrant workforce, upskilling and cross-skilling in-house staff, hiring gig workers, or even joining hands with startups and accelerators.

Robust education programs must also be created in areas where the fabs are being built, starting even at the K-12 level, said Sergey Shchemelev, a managing director at Deloitte Consulting and the firm’s U.S. human capital semiconductor leader.

Shchemelev noted that the semiconductor industry has not been especially attractive and “Especially here in the U.S., it’s definitely not one that’s been growing; in fact, it’s declining, so now we’re trying to get folks interested in it, especially younger folks coming out of high school, trade school, and community colleges.”

One of the problems is a “lack of transparency into the career path” of working at a fab plant, he said. These companies not only need to promote what a clear career path looks like, but also company values, vision, and culture. That will make it “a much more clear and attractive pathway for someone to join” the industry, Shchemelev said. Those are known quantities at big tech companies like Google and Microsoft, he said, “So having that clarity into how you’re going to progress and how you can grow your skillset, and how transferrable your skillset is and compensation, will make [semiconductor jobs] attractive.”

Upskilling is not relevant for jobs in the semiconductor industry, maintained Gaurav Gupta, a vice president and analyst at Gartner. “That hardly works; to some extent, there is continuous upskilling, but in a lot of cases for emerging and new technologies, companies look for new hires.”

For workers who do want to upskill, however, the focus should be on “combining a foundational understanding of semiconductor technology and circuits with expertise in an application area they’re passionate about,’’ said Paul Farnsworth, CTO of DHI Group, parent company of tech job site Dice. “For example, if you’re interested in biomedical applications, integrating that focus with knowledge of semiconductors can open doors to innovation in electronics for healthcare.”

Since semiconductors power technologies like AI, high-speed communication, and integrated circuit design, building skills in related fields such as mathematics, signal processing, and software is essential, Farnsworth said. “Developing comfort with hardware and electronics, alongside software proficiency, is becoming foundational for advancing in this dynamic field. Ultimately, aligning your upskilling efforts with personal interests while mastering these core technical areas creates a strong platform for future growth.”

It’s an uphill battle trying to attract talent for chip manufacturing in the U.S. where the fabs have to compete with the big-name tech companies, Gupta said. Contrast that with TSMC and Samsung in Taiwan and South Korea, where “they are the premiere companies. But in the U.S., that will always be a struggle.”

Mark Granahan, cofounder and CEO of iDEAL Semiconductor, agreed, saying the challenge to find workers is greater because of the CHIPS Act. iDEAL is in the midst of qualifying its technology and that requires a lot of reliability testing, Granahan said. A college degree is not required for this role; the company is looking for people who are “logical in their workflows and [can] keep good records and be consistent in what they do,’’ he said, adding that “We’re always hiring.”

To find people, Granahan has formed partnerships with search firms across the country, and also relies on recruitment platforms like Handshake, which is mainly used by college students and recent graduates.

He remains optimistic that the tide is turning for the semiconductor industry. “One of the great side effects of the CHIPS Act is … the U.S. government has large megaphones. There’s a lot of interest out there in semiconductors due to the criticality that’s been communicated by the federal government [about] our nation’s ability to produce semiconductors. It’s a critical piece of infrastructure we need, so fundamentally, I think … that’s driving interest.”

Esther Shein is a freelance technology and business writer based in the Boston area.

]]>
https://cacm.acm.org/news/staffing-for-semiconductors/feed/ 0 764517
Shaping the Future of Technology https://cacm.acm.org/blogcacm/shaping-the-future-of-technology/ https://cacm.acm.org/blogcacm/shaping-the-future-of-technology/#respond Wed, 29 Jan 2025 21:31:32 +0000 https://cacm.acm.org/?p=764779 In the ever-evolving world of technology, the influence of computer science on society is both profound and far-reaching. Once a niche field reserved for specialized researchers, computing has transformed into an essential discipline that shapes nearly every aspect of modern life. With the rise of online education and the growing accessibility of computer science degrees, we are entering an era where more people than ever can contribute to this transformative field.

However, alongside these advancements come critical ethical questions that demand our attention. We gain valuable insights by examining the intersection of online computer science education, computing’s impact on society, and the ethical implications of these developments. These insights help us better understand how to shape technology’s future responsibly.

The Rise of Online Computer Science Education

As higher education costs continue to rise, many learners seek flexible, budget-friendly alternatives that still provide high-quality education. This is especially true for technical fields like computer science, where affordability and accessibility can make a significant difference.

Over the past decade, one of the most significant shifts in education has been the proliferation of online learning platforms. The availability of cheap computer science degree online programs has democratized access to one of the most in-demand fields of study. As technology becomes more pervasive, the demand for skilled professionals who can understand and innovate responsibly has skyrocketed. Online programs have become a gateway for many to achieve this.

Pursuing a computer science degree online enables students to continue working or managing other responsibilities while studying. Many of these programs provide flexible schedules, allowing learners to pace their education according to their needs without compromising quality.

Benefits of Online Education in Computer Science

Flexibility

Students can balance their studies with work, family commitments, or other personal responsibilities, making it an ideal choice for diverse learners. Additionally, online programs often offer various course options and specializations. That allows students to tailor their education to their specific career goals. This flexibility can lead to a more personalized and rewarding learning experience for individuals pursuing a computer science education.

Accessibility

Non-traditional students, including career changers and those in remote regions, can access programs that were once out of reach. Students can study from anywhere with an internet connection, saving time and money on commuting. Overall, online programs provide a convenient and customizable option for computer science students.

Affordability

Lower financial barriers make education attainable for a broader demographic, especially with online computer science degrees that offer substantial value. Flexibility in scheduling and the opportunity to continue working while pursuing a degree are additional benefits of online programs. These advantages make them an attractive option for many prospective students.

However, this democratization of education comes with challenges:

Quality Concerns

Can online programs adequately prepare students for industry demands, especially in fields requiring hands-on problem-solving? Additionally, the lack of face-to-face interaction with professors and peers may hinder some students’ learning experiences. It is important for online programs to continuously adapt and improve their curriculum to address these potential drawbacks.

Ethical Preparedness

Are graduates equipped to navigate the ethical complexities of computing, particularly in high-stakes industries? Online programs must prioritize ethics education to ensure graduates have the necessary skills to make career decisions. Incorporating real-world case studies and interactive discussions can help students develop a strong ethical foundation.

Computing’s Impact on Society

Computing technologies have become the backbone of the modern world, transforming industries and daily life. From the automation of repetitive tasks to groundbreaking AI innovations, technology is reshaping society at an unprecedented pace.

In 2024, 6.4% of students in 41 U.S. states were enrolled in foundational computer science courses, according to research. This indicates a growing recognition of the importance of computer science education in preparing students for the future job market. As technology advances, institutions must prioritize teaching ethical considerations and practical problem-solving skills in computing.

As computing continues to advance, it is crucial for educational programs to address these challenges. They must ensure that graduates are technically proficient and ethically responsible. Integrating ethics into computer science curriculums can help students understand the implications of their work on society. It also enables them to make informed decisions in their future careers.

Examples of Computing’s Influence

Artificial Intelligence (AI)

It powers applications such as facial recognition and predictive analytics. This raises concerns about privacy, bias, and accountability in decision-making.

Social Media Algorithms

Redefine human interaction by curating content and connecting users all over the world. Contribute to the spread of misinformation and polarization.

Positive Impacts of Computing

Healthcare Advancements

AI-driven diagnostics and wearable health monitors save lives by enabling early detection and personalized care. Using computing in healthcare has also improved patient outcomes, reduced medical errors, and increased efficiency in healthcare delivery. Overall, the positive impacts of computing on healthcare are undeniable and continue to revolutionize the industry.

Educational Access

E-learning platforms break geographical barriers, providing global knowledge-sharing and career advancement opportunities. Furthermore, integrating computing into education has allowed for personalized learning experiences and increased access to resources for all students. Overall, computing has dramatically expanded educational opportunities and improved learning outcomes on a global scale.

Sustainability

Technologies that optimize energy use, improve resource efficiency, and support renewable energy adoption pave the way for a greener future. These technologies benefit the environment and offer cost-saving solutions for businesses and individuals. Investing in sustainability technologies can create a more sustainable and environmentally friendly world for future generations.

Challenges and Consequences

Automation and Job Displacement

Economic inequality arises in disrupted sectors as specific roles become obsolete. This can lead to job displacement and financial instability for those affected. Governments and businesses must address these challenges by providing retraining programs and support for displaced workers to ensure a smooth transition to new opportunities.

Environmental Costs

Data centers and cryptocurrency mining consume vast amounts of energy, contributing to environmental degradation. This increased energy consumption can exacerbate climate change and harm ecosystems. Industries must adopt sustainable practices and invest in renewable energy sources to mitigate these environmental costs.

Ethical Implications of Computing

The ethical challenges of computing are as diverse as its applications. As technology integrates further into daily life, the stakes for ethical oversight continue to rise.

Data Privacy

Personal information is critical to business models but susceptible to breaches and misuse. Online computer science programs must teach privacy protection methods and develop systems that prioritize user trust.

Algorithmic Bias

AI systems can perpetuate societal biases if trained on flawed or incomplete data. Addressing algorithmic fairness necessitates technical expertise, ethical awareness, and the ability to envision inclusive systems.

Automation’s Workforce Impact

Efficiency gains from automation disrupt traditional industries, displacing workers and creating economic disparities. Developers must balance innovation and society’s need for job creation and retraining.

Integrating Ethics into Online Computer Science Education

As demand for cheap computer science degrees online grows, educational institutions must emphasize ethics as a foundational curriculum component. This will ensure that future technologists understand the importance of considering ethical implications. By integrating ethics into online computer science education, we can help shape a more responsible and socially conscious tech industry. This goes beyond technical training to include the broader societal context in which technology operates.

Case Studies

Analyze real-world ethical dilemmas, such as AI controversies or data breaches, to develop practical problem-solving skills. Discussions around high-profile cases can help students understand the implications of their work.

Interdisciplinary Learning

Explore intersections with sociology, psychology, and law to foster a holistic understanding of societal impacts. Encourage students to think critically about the role of technology in diverse cultural and economic contexts.

Cultivating Responsibility

Create opportunities for peer discussions, mentorships, and collaborations with industry professionals. Adapt curriculums to reflect emerging technologies and the ethical questions they raise.

The Role of the Broader Computing Community

Addressing the ethical implications of computing requires collaboration across sectors, uniting stakeholders to establish best practices and shared values. Computational science is a rapidly evolving field requiring ongoing dialogue and cooperation to ensure responsible innovation and decision-making.

Key Stakeholders

Industry Leaders

Adopt transparent practices and prioritize ethical product development. Implement bias testing and accountability mechanisms in technologies such as AI.

Policymakers

Establish regulations that protect users and encourage responsible innovation without stifling creativity. Collaborate with industry leaders and researchers to stay informed on technological advancements and potential risks.

Researchers

Study societal impacts and propose solutions to emerging challenges. Bridge the gap between academic theory and real-world application through partnerships with industry.

Shaping a Responsible Future

Online computer science degrees bring fresh perspectives to the field, allowing individuals from diverse backgrounds to contribute to technological innovation. This democratization of education opens doors to new ideas, but it also emphasizes the need for shared responsibility among educational institutions, industry leaders, and policymakers. Together, they must ensure that ethics remain a central focus, not an afterthought, in computing.

Preparing for the challenges of tomorrow requires fostering interdisciplinary collaboration and prioritizing responsible practices. Integrating ethical frameworks into educational curriculums and professional standards can shape a future where technology serves humanity’s best interests. With thoughtful planning and cooperative efforts, we can ensure that our choices today lead to a more equitable, sustainable, and innovative technological landscape for future generations.

Alex Tray is a system administrator and cybersecurity consultant with 10 years of experience. He is currently self-employed as a cybersecurity consultant and as a freelance writer.

]]>
https://cacm.acm.org/blogcacm/shaping-the-future-of-technology/feed/ 0 764779