When Research Ignores Consent: Why Non-Consensual Web Scanning Fails Academic Ethics

When Research Ignores Consent: Why Non-Consensual Web Scanning Fails Academic Ethics

This article examines a University of Georgia web scanning project that probes servers globally without prior consent. It challenges the claim that "educational purpose" justifies non-consensual data collection, dismantles the opt-out-as-consent fallacy, and argues that academia must meet the same ethical standards society demands of corporations. If consent doesn't come first, it isn't ethics — it's convenience.

Introduction: A Real Case, Not a Hypothetical

In recent years, academic institutions have increasingly engaged in large-scale internet measurement research. One such case is a publicly documented web and TLS scanning project conducted under the name of the University of Georgia, explained through a scan-notification page intended to justify unsolicited connections made to servers across the internet.

The researchers describe their activity as educational, non-intrusive, and conducted in the interest of science. Yet a fundamental ethical question remains unresolved:

Can research be ethical if it is conducted without prior consent, even when labeled “educational”?

This article argues that it cannot.


Who Is Involved and What Is Being Done

According to the public scan notice, the project is affiliated with UGA’s School of Computing and lists specific individuals responsible for the activity, including Roberto Perdisci and graduate researcher Xingda Bao.

The activity involves automated scanning of publicly reachable servers across the internet, collecting information such as TLS certificates, HTTP headers, and landing-page content. The stated goal is to analyze the security ecosystem and identify trends related to certificates and potential misuse.

The researchers emphasize that:

  • Only publicly accessible data is collected

  • No exploitation is intended

  • An opt-out mechanism is available

None of these claims resolve the core ethical issue: participation is imposed, not agreed to.


The Central Ethical Issue: No Prior Consent

The defining problem is not the technology, nor even the data itself. It is the absence of permission.

Website owners are included automatically, without notice, consent, or agreement. Public accessibility is treated as implicit approval for automated research.

That assumption is ethically unsound.

Public exposure does not imply consent. Silence does not imply consent. Being reachable does not imply permission to be studied, catalogued, or analyzed at scale.


Why “Opt-Out” Is Not Consent

The scan notice provides an opt-out option: if you do not want to be scanned, you may contact the researchers and request exclusion.

This is not consent. It is retroactive damage control.

Ethical consent must be prior, informed, and voluntary. Opt-out fails on all three fronts. The scan has already occurred. The burden shifts entirely to the unaware target to discover the practice and object.

This is not a choice — it is a procedural afterthought.


“Educational Purpose” Does Not Make an Act Ethical

The primary defense offered is education.

But explanation does not equal authorization.

If stating “this is for education” were sufficient, then logically:

  • A chemistry lecturer could manufacture illegal drugs “for teaching”

  • A horticulture instructor could grow marijuana “for educational research”

  • A researcher could bypass safeguards simply by publishing a justification page afterward

In every one of these cases, society already understands the boundary:

Intent does not override consent, law, or ethics.

Education demands higher moral standards, not exceptions.


There Are Ethical Alternatives — and They Are Easy

If the purpose is genuinely educational or methodological, there is no necessity to involve uninformed, non-consenting parties worldwide.

Ethical alternatives are readily available:

  • Dummy websites can be created in minutes

  • Synthetic environments can be generated using cloud platforms

  • AI agents can produce realistic but non-real datasets

  • Universities already operate their own domains and infrastructure

Researchers can:

  • Scrape intentionally created test sites

  • Use school-owned systems

  • Work with explicitly consenting participants

When non-consensual global scanning is chosen despite these options, the issue is no longer technical feasibility. It is choice.


Why Global, Non-Consensual Scraping Is Indefensible

This practice fails a basic moral test:

  • The subjects did not agree

  • The subjects receive no benefit

  • The subjects bear the operational and security risk

  • The researchers bear none of the cost

That is not collaboration. It is extraction.

Taking data without permission — when permission could reasonably be obtained or avoided — is morally indistinguishable from stealing, regardless of motive.


If Cambridge Analytica Was Wrong, This Should Concern You Too

Readers should ask themselves a simple question:

If the Cambridge Analytica scandal was unethical, why is this different?

In that case, data was collected and repurposed without meaningful user consent. The public reaction was clear: lack of consent mattered more than claimed intent.

The same was true in the Facebook data misuse scandals. Users were told data was “public” or “available,” yet society overwhelmingly rejected the idea that availability equals permission.

The pattern is the same:

  • Data collected without consent

  • Justified after the fact

  • Responsibility shifted to the affected party

If we condemn those practices in industry, we cannot excuse them in academia simply because the word “research” is used.

Ethical standards must be consistent — or they are meaningless.


Institutional Trust and Misuse of Resources

Universities benefit from extraordinary public trust. Their names, IP ranges, and infrastructure signal legitimacy and ethical oversight.

When non-consensual scanning is conducted under an academic banner, the institution implicitly endorses — or at least tolerates — the method.

Universities are not neutral platforms. They exist to model integrity and responsibility.

Normalizing “take first, explain later” undercuts that mission.


Lessons From Medical Research Ethics

History shows where this logic leads.

Medical research once justified unethical actions “for science.” Society responded by establishing consent requirements, oversight bodies, and ethical review frameworks.

The lesson was simple and hard-earned:

Good intentions are not enough.

Researchers cannot be the sole judges of what is acceptable.


COVID-19 and Why Oversight Still Matters

There is no conclusive evidence that COVID-19 was deliberately engineered or intentionally released. The World Health Organization has stated that the origins of COVID-19 remain under investigation.

But intent is not the point.

Even accidental harm demands accountability, safeguards, and oversight. The absence of malice does not remove ethical responsibility.


When Research Methods Resemble Threat Activity

From the target’s perspective, automated probing without consent looks identical to hostile reconnaissance.

Ethics are judged by actions, not titles.

If research methods resemble those of threat actors, trust collapses — regardless of the justification offered afterward.


What Ethical Internet Research Should Look Like

Ethical research is harder and more limited — and that is its strength.

It requires:

  • Consent-first approaches

  • Advance transparency

  • Respect for refusal by default

  • Acceptance that not all data should be collected

Ethical limits are not obstacles. They are safeguards.


Conclusion: Ethics Before Convenience

Education does not justify taking without asking.

Consent obtained after the fact is not consent. Transparency without permission is not ethics — it is public relations.

If society condemns non-consensual data collection by corporations, it must hold academia to the same standard.

If research cannot be conducted ethically, then it should not be conducted at all.


"Why We Wrote This" We discovered this scanning activity firsthand — it was detected targeting a user's website hosted on an s͛Card custom domain. Because s͛Card custom domains route through our infrastructure, we're able to monitor and flag suspicious activity like unauthorized academic scraping on behalf of our users. This added layer of protection is built into every s͛Card custom domain, giving users visibility and defense against threats they might never notice on their own.

Share
Eric Lim

Creative Director

Eric Lim
I'm an observer by training, equipped with Viscom + Design + Craft + Technology skillsets. You can talk to me if you need business solutions. I would love to help you. Lastly, thank you very much for visiting my profile, and I hope you'll love to use s͛Card, which is built by my team at DMW+SI.
scard-userelsg

Other Articles By Eric Lim

news letter

Get the most popular topic straight to your inbox!

Every month, our expert team sifts through tech, culture and business news to bring you the most pertinent information for our engaged readership of thousands.