Discussion:
[tor-dev] User perception of onion service discovery
Philipp Winter
2017-10-04 00:25:15 UTC
Permalink
This is the second part of our preliminary analysis of how Tor users
interact with onion services [0]. In this part, we analyse the issue of
onion service discovery. Onion services are private by default, so it's
the operator's responsibility to disseminate their domain if they want
it to be public.
How do you discover new onion sites?
The breakdown looks as follows. Respondents could select multiple
answers, so the percentages are based on the total number of
respondents.

Method Percentage
-----------------------------------------------------------------------
From social networking sites such as Reddit or Twitter 50.67
I browse the list of onion site search engines such as ahmia.fi 50.50
I randomly encounter them while browsing the web 47.65
Recommendations from friends and family 19.46
Other (see below for common themes) 18.12
I am not interested in learning about new onion sites 4.19

The data shows that social networking sites, search engines, and "random
encounters" are rather popular. Respondents who selected "Other" mostly
brought up onion service lists and aggregators.
Are you satisfied with the way you discover new onion sites?
61% selected "Yes" while the remaining 39% selected "No."

Some respondents who selected "Yes" brought up that they have no
interest in learning about new onion services; in part because they
only use Facebook's (or some other) onion service.

Among the respondents who selected "No," there are a bunch of
reoccurring themes, in no particular order:

- The most prominent complaint was about broken links on onion site
lists. There is non-trivial churn among onion sites and our
respondents were frustrated that existing lists are typically not
curated and contain many dead links.

- Many respondents were not aware of search engines such as ahmia.fi.
Among those that were, many were not satisfied with both the search
results and the number of indexed onion sites. Unsurprisingly,
a "Google for onion sites" was a frequent wish.

- Several respondents were unhappy with existing aggregators. In
addition to broken links, some distrust lists because they
occasionally contain scam and phishing sites. The difficulty of
telling apart two given onion domain names exacerbates this issue.

- Some respondents would like aggregators to be more verbose in their
description of onion sites. In particular, these respondents were
trying to avoid illegal and pornographic content, which is often
difficult if the description is vague and the onion domain reveals
nothing about its content.

- Many respondents expressed frustration about the difficulty of finding
out if site X also provides a corresponding onion service. A common
wish was to have site X list its onion service prominently in a footer.
Ironically, some respondents were surprised that torproject.org has a
corresponding onion site -- they couldn't find it on the web site.

- Two respondents compared the current state of onion services with the
web of the 90s: Few sites existed, they linked to each other only
sparsely, and search engines were experimental at best.

- Interestingly, some respondents voiced frustration about various
usability issues, but mentioned in the same sentence that this is an
inherent trade-off of privacy technology, suggesting that there is
nothing that can be done about it.

There are two potential solutions that would address some of the above
issues:

- Have next-gen onion services opt-in to a broadcast mechanism that
automatically propagates them. Naturally, we would like such a
mechanism to be censorship-resistant and built in a way that only the
owner of an onion service is authorised to broadcast their service.

- Websites could use an HTTP header to announce the existence of a
corresponding onion site. This issue was discussed in Feb 2017 over
at tor-onions. Someone brought up the Alt-Svc header as a potential
solution [1]. In a subsequent survey question we asked if our
respondents would appreciate an automatic redirect from a web site to
its corresponding onion site. The overall tendency leaned towards
"Yes," provided that the implementation is sound and users can
override the redirect.

Again, it's important to take these results with a grain of salt. Our
data has some survivor bias: Presumably, we mostly heard from people who
are Tor users despite usability issues. We likely didn't hear from many
people who once experimented with Tor or onion services, decided it's
not usable enough, and gave up.

The above was joint work with my colleagues Marshini Chetty, Annie
Edmundson, Nick Feamster, and Laura M. Roberts.

[0] <https://nymity.ch/onion-services/>
[1] <https://lists.torproject.org/pipermail/tor-onions/2016-February/000045.html>
Philipp Winter
2017-10-14 14:45:18 UTC
Permalink
Post by Philipp Winter
- Many respondents were not aware of search engines such as ahmia.fi.
Among those that were, many were not satisfied with both the search
results and the number of indexed onion sites. Unsurprisingly,
a "Google for onion sites" was a frequent wish.
Someone at the Tor dev meeting brought up that Google is indexing
tor2web.org including all the onion sites it knows about. That
means that we can use Google as a search engine for (a subset of)
onion services by searching for "site:onion.to foo".
dawuud
2017-10-14 15:12:05 UTC
Permalink
That sounds terrabad. Can we finally set fire to tor2web? It was
never a good idea.
Post by Philipp Winter
Post by Philipp Winter
- Many respondents were not aware of search engines such as ahmia.fi.
Among those that were, many were not satisfied with both the search
results and the number of indexed onion sites. Unsurprisingly,
a "Google for onion sites" was a frequent wish.
Someone at the Tor dev meeting brought up that Google is indexing
tor2web.org including all the onion sites it knows about. That
means that we can use Google as a search engine for (a subset of)
onion services by searching for "site:onion.to foo".
_______________________________________________
tor-dev mailing list
https://lists.torproject.org/cgi-bin/mailman/listinfo/tor-dev
Andreas Krey
2017-10-14 16:50:17 UTC
Permalink
Post by dawuud
That sounds terrabad. Can we finally set fire to tor2web? It was
never a good idea.
Why? There is exactly nothing that would stop google from actually
indexing .onion domains (it knows about), like it now is with
onion.to, even thought that happens a) accidentally and b) requires
there being links to .onion.to instead of to .onion on the clearweb.

- Andreas
--
"Totally trivial. Famous last words."
From: Linus Torvalds <torvalds@*.org>
Date: Fri, 22 Jan 2010 07:29:21 -0800
dawuud
2017-10-14 18:43:03 UTC
Permalink
Plaintext communications intermediaries like tor2web violate the end
to end principle and the principle of least authority. If we as the
Tor community are committed to human rights then it follows we would
abolish terrible things like tor2web or at least frown upon it's use.

I am frowning so hard right now.


Sincerely,

David Stainton
Post by Andreas Krey
Post by dawuud
That sounds terrabad. Can we finally set fire to tor2web? It was
never a good idea.
Why? There is exactly nothing that would stop google from actually
indexing .onion domains (it knows about), like it now is with
onion.to, even thought that happens a) accidentally and b) requires
there being links to .onion.to instead of to .onion on the clearweb.
- Andreas
--
"Totally trivial. Famous last words."
Date: Fri, 22 Jan 2010 07:29:21 -0800
_______________________________________________
tor-dev mailing list
https://lists.torproject.org/cgi-bin/mailman/listinfo/tor-dev
Alec Muffett
2017-10-15 08:08:14 UTC
Permalink
Post by dawuud
Plaintext communications intermediaries like tor2web violate the end
to end principle and the principle of least authority. If we as the
Tor community are committed to human rights then it follows we would
abolish terrible things like tor2web or at least frown upon it's use.
I would recommend continuing to enable/support Tor2Web, or at least not
moving to make such a solution inoperable.


Dawuud is absolutely right re: violation of E2E* and a bunch of other
criticisms also apply; however I have three observations on this topic:

1) Someone invented Tor2web, therefore someone else is likely to want to
reimplement it; ideas tend to persist in this way

2) (as observed above) Google *do* crawl onion sites via "onion.to", which
is a fun surprise for people who insist that "The Dark Web Is Not Indexed
And Is Therefore Spooky"

3) Making such a move to block Tor2web-like sites might engender false
trust amongst the people who set up Onion sites: "It's Okay, Google Can't
Get At Us"


I would recommend investing more effort in Tor2web/similar, because having
a permeable barrier between IP-Space and OnionSpace appears useful.

At very most I might propose that:

a) OnionSites become aware of the X-Tor2web header which (from legit T2W
instances, at least) permits the OnionSite operator to block or redirect
the user to use a "proper" Onion network connection

b) That TheTorProject consider indexing known Tor2web sites and publish
them, perhaps adding a feature to optionally block them from TorBrowser
access**, thereby to prevent stupid intra-Tor deanonymisation loops

- a


*although speaking as a geek I believe that re-engineering T2W to support
SSL via SNI-Sniffing would address this, it would be a gross and pointless
hack, complicated still further by certificate issuance, and all reasonable
use cases for which would be better addressed by running a local copy of
Tor.

**the hardcore alternative of blocking them from being accessed by exit
nodes causing a likely-intolerable argument.
--
http://dropsafe.crypticide.com/aboutalecm
teor
2017-10-15 12:03:27 UTC
Permalink
Post by dawuud
Plaintext communications intermediaries like tor2web violate the end
to end principle and the principle of least authority. If we as the
Tor community are committed to human rights then it follows we would
abolish terrible things like tor2web or at least frown upon it's use.
I would recommend continuing to enable/support Tor2Web, or at least not moving to make such a solution inoperable.
v2 onion service Tor2web would be easy for HSDirs to block, due to an
implementation bug. We've chosen not to block it. But we haven't spent
much time on fixing its bugs, either.

As far as I am aware, no-one is writing Tor2web for v3 onion services.

We have open tickets for protecting relays that handle onion service traffic
from knowing both the client and service IP address.

So if anyone does write v3 Tor2web, they will need to write it so it:
* uses a 3-hop path for all descriptors, because otherwise that can be used
for a selective denial of service;
* uses a 3-hop path to connect to intro and rend when a descriptor has the
single onion service flag;
* retry using a 3-hop path on failure (internal reachability or actual connection
failure)

And I'm not sure whether we would merge this feature into core tor, due to the
user security issues that David and others have mentioned.

T
dawuud
2017-10-15 17:18:18 UTC
Permalink
I agree with Alec. Don't block the existing tor2web stuff,
that would be very rude. Instead just do not implement any
kind of tor2web for v3 onion services so that tor2web will
gradually fade as we migrate.
Post by Alec Muffett
*although speaking as a geek I believe that re-engineering T2W to
support SSL via SNI-Sniffing would address this, it would be a gross
and pointless hack, complicated still further by certificate issuance,
and all reasonable use cases for which would be better addressed by
running a local copy of Tor.
Ah yeah, Donncha wrote a tool to do that called oniongateway:

https://github.com/DonnchaC/oniongateway

Is that what you mean?

Loading...