DeepWeb Ad

Saturday, January 24, 2015

About Deep Web

Deep Web (also called the Deepnet,Invisible Web, or Hidden Web) is the portion of World Wide Web content that is not indexed by standard search engines.
Mike Bergman, founder of BrightPlanet and credited with coining the phrase, said that searching on the Internet today can be compared to dragging a net across the surface of the ocean: a great deal may be caught in the net, but there is a wealth of information that is deep and therefore missed. Most of the Web's information is buried far down on sites, and standard search engines do not find it. Traditional search engines cannot see or retrieve content in the deep Web. The portion of the Web that is indexed by standard search engines is known as the Surface Web. As of 2001, the deep Web was several orders of magnitude larger than the surface Web.
The deep web should not be confused with the dark Internet, computers that can no longer be reached via the Internet. The Darknet distributed file sharing network, can be classified as part of the Deep Web.
Although much of the Deep Web is innocuous, some prosecutors and government agencies, among others, are concerned that the Deep Web is a haven for serious criminality.

Size

Bright Planet, a web-services company, describes the size of the Deep Web in this way:
It is impossible to measure or put estimates onto the size of the deep web because the majority of the information is hidden or locked inside databases. Early estimates suggested that the deep web is 400 to 550 times larger than the surface web. However, since more information and sites are always being added, it can be assumed that the deep web is growing exponentially at a rate that cannot be quantified. Estimates based on extrapolations from a study done at University of California, Berkeley in 2001 speculate that the deep web consists of about 7.5 petabytes. More accurate estimates are available for the number of resources in the deep Web: research of He et al. detected around 300,000 deep web sites in the entire Web in 2004, and, according to Shestakov, around 14,000 deep web sites existed in the Russian part of the Web in 2006.

Naming

Bergman, in a seminal paper on the deep Web published in The Journal of Electronic Publishing, mentioned that Jill Ellsworth used the term invisible Web in 1994 to refer to websites that were not registered with any search engine. Bergman cited a January 1996 article by Frank Garcia:
It would be a site that's possibly reasonably designed, but they didn't bother to register it with any of the search engines. So, no one can find them! You're hidden. I call that the invisible Web.
Another early use of the term Invisible Web was by Bruce Mount and Matthew B. Koll of Personal Library Software, in a description of the @1 deep Web tool found in a December 1996 press release.
The first use of the specific term Deep Web, now generally accepted, occurred in the aforementioned 2001 Bergman study.

Methods

Methods which prevent web pages from being indexed by traditional search engines may be categorized as one or more of the following:
  • Dynamic content: dynamic pages which are returned in response to a submitted query or accessed only through a form, especially if open-domain input elements (such as text fields) are used; such fields are hard to navigate without domain knowledge.
  • Unlinked content: pages which are not linked to by other pages, which may prevent Web crawling programs from accessing the content. This content is referred to as pages without backlinks (also known as inlinks). Also, search engines do not always detect all backlinks from searched web pages.
  • Private Web: sites that require registration and login (password-protected resources).
  • Contextual Web: pages with content varying for different access contexts (e.g., ranges of client IP addresses or previous navigation sequence).
  • Limited access content: sites that limit access to their pages in a technical way (e.g., using the Robots Exclusion Standard or CAPTCHAs, or no-store directive which prohibit search engines from browsing them and creating cached copies.)
  • Scripted content: pages that are only accessible through links produced by JavaScript as well as content dynamically downloaded from Web servers via Flash or Ajax solutions.
  • Non-HTML/text content: textual content encoded in multimedia (image or video) files or specific file formats not handled by search engines.
  • Software: Certain content is intentionally hidden from the regular internet, accessible only with special software, such as Tor. Tor allows users to access websites using the .onion host suffix anonymously, hiding their IP address. Other such software includes I2P and darknet software.

Indexing the Deep Web

While it is not always possible to directly discover a specific web server's content so that it may be indexed, a site potentially can be accessed indirectly (due to computer vulnerabilities).
To discover content on the Web, search engines use web crawlers that follow hyperlinks through known protocol virtual port numbers. This technique is ideal for discovering content on the surface Web but is often ineffective at finding Deep Web content. For example, these crawlers do not attempt to find dynamic pages that are the result of database queries due to the indeterminate number of queries that are possible. It has been noted that this can be (partially) overcome by providing links to query results, but this could unintentionally inflate the popularity for a member of the deep Web.
DeepPeep, Intute, Deep Web Technologies, Scirus, and Ahmia.fi are a few search engines that have accessed the Deep Web. Intute ran out of funding and is now a temporary static archive as of July, 2011. Scirus retired near the end of January, 2013.
Researchers have been exploring how the Deep Web can be crawled in an automatic fashion, including content that can be accessed only by special software such as Tor. In 2001, Sriram Raghavan and Hector Garcia-Molina (Stanford Computer Science Department, Stanford University) presented an architectural model for a hidden-Web crawler that used key terms provided by users or collected from the query interfaces to query a Web form and crawl the Deep Web content. Alexandros Ntoulas, Petros Zerfos, and Junghoo Cho of UCLA created a hidden-Web crawler that automatically generated meaningful queries to issue against search forms. Several form query languages (e.g., DEQUEL) have been proposed that, besides issuing a query, also allow extraction of structured data from result pages. Another effort is DeepPeep, a project of the University of Utah sponsored by the National Science Foundation, which gathered hidden-Web sources (Web forms) in different domains based on novel focused crawler techniques.
Commercial search engines have begun exploring alternative methods to crawl the deep Web. The Sitemap Protocol (first developed, and introduced by Google in 2005) and mod oai are mechanisms that allow search engines and other interested parties to discover deep Web resources on particular Web servers. Both mechanisms allow Web servers to advertise the URLs that are accessible on them, thereby allowing automatic discovery of resources that are not directly linked to the surface Web. Google's deep Web surfacing system pre-computes submissions for each HTML form and adds the resulting HTML pages into the Google search engine index. The surfaced results account for a thousand queries per second to deep Web content. In this system, the pre-computation of submissions is done using three algorithms:
  1. selecting input values for text search inputs that accept keywords,
  2. identifying inputs which accept only values of a specific type (e.g., date), and
  3. selecting a small number of input combinations that generate URLs suitable for inclusion into the Web search index.
In 2008, to facilitate users of Tor hidden services in their access and search of a hidden .onion suffix, Aaron Swartz designed Tor2web—a proxy application able to provide access by means of common web browsers. Using this application, Deep Web links appear as a random string of letters followed by the .onion TLD. For example, http://xmh57jrzrnw6insl followed by .onion, links to TORCH, the Tor search engine web page.

Classifying resources

Most of the work of classifying search results has been in categorizing the surface Web by topic. For classification of deep Web resources, Ipeirotis et al. presented an algorithm that classifies a deep Web site into the category that generates the largest number of hits for some carefully selected, topically-focused queries. Deep Web directories under development include OAIster at the University of Michigan, Intute at the University of Manchester, Infomine at the University of California at Riverside, and DirectSearch (by Gary Price). This classification poses a challenge while searching the deep Web whereby two levels of categorization are required. The first level is to categorize sites into vertical topics (e.g., health, travel, automobiles) and sub-topics according to the nature of the content underlying their databases.
The more difficult challenge is to categorize and map the information extracted from multiple deep Web sources according to end-user needs. Deep Web search reports cannot display URLs like traditional search reports. End users expect their search tools to not only find what they are looking for, but to be intuitive and user-friendly. In order to be meaningful, the search reports have to offer some depth to the nature of content that underlie the sources or else the end-user will be lost in the sea of URLs that do not indicate what content lies beneath them. The format in which search results are to be presented varies widely by the particular topic of the search and the type of content being exposed. The challenge is to find and map similar data elements from multiple disparate sources so that search results may be exposed in a unified format on the search report irrespective of their source.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

Friday, January 23, 2015

Deep Web Marketplaces

Over the past couple of weeks, I’ve been frequenting the deep web marketplaces most famously used for buying drugs online with Bitcoin.
I wanted to see if there was anything we could learn about how these illicit marketplaces work that could be applied to improve the legal marketplaces we invest in at USV.
As part of my research, I purchased an item on Evolution (no, not drugs – a pair of furry boots) in an effort to understand the dynamics of these marketplaces, from trust and safety to flow of funds. This is what I learned in the process.
Privacy
  • Deep web marketplaces can only be accessed using Tor, a decentralized computer network that anonymizes traffic such that it’s harder to trace an individual user through their IP address. If you’d like to learn more about Tor and how it works, this is a good introduction.
  • Most if not all marketplaces force you to sign up before browsing the listings. The sign up process involves picking a username and password, and an account PIN number. You’re also expected to remember a mnemonic private key for your account, which is not stored on the marketplace’s servers.
  • Most sellers (particularly drug dealers) require all communications to be encrypted with PGP. Most marketplaces have a PGP key field at the profile setup level.
  • Some marketplaces automatically delete all order information from their servers 30 days after an order has been “finalized” by the user.
  • No e-mails are used, only Bitmessage (decentralized) or the marketplace’s messaging system (encrypted and periodically deleted).
Products
  • Lots of drugs. The drugs category is 10x larger than all others. You’ll find anything from Valium to cocaine and LSD.
  • You’ll also find digital content, stolen credit card and user/password lists, hacking services (mostly DDoS), counterfeit goods (fashion, jewelry, etc.), lab equipment, electronics (I was tempted to buy a pocket-sized EMP pulse generator), high-end spy gear, forged documents (driver’s licenses, passports), counterfeit currency, weapons and more.
Brand and Reputation
  • Brand and reputation means everything to sellers. Buyers guide themselves via eBay-style reviews of the sellers.
  • This is particularly important in an environment where there is no real identity shared between any of the participants. By contrast, I may not know who an eBay seller is but I take comfort in knowing that eBay does.
  • Most sellers have 95%+ positive ratings. Some sellers have been involved in over 10,000 transactions.
  • Many sellers have a presence across multiple deep web marketplaces, and oftentimes point to their profiles on different platforms as a way to further establish credibility.
  • The community moderates sellers beyond the eBay-style reviews. A lot of marketplaces have separate community forums where users review sellers and products.
  • A quick way for new sellers to establish credibility is to get reviewed by these community members.
  • These forums often have established members, to whom sellers frequently send review samples.
  • Sellers oftentimes link to these reviews as social proof, which are often rich in detail about the quality of the product (with pictures!), the seller, the packaging (good/bad stealth), speed, etc.
Flow of funds
  • You are given a bitcoin public key on to which you must deposit funds before making a purchase.
  • You’d buy bitcoin at an exchange, and use a mixing/tumbling service to anonymize them for a small fee. The need for these services in illegal transactions is interesting, since Bitcoin is frequently antagonized for its anonymity.
  • You are expected to trust the marketplace with holding your funds. Some users keep a balance on their account, while others only make a deposit when they intend to make a purchase.
  • Funds show up in your account once the transaction has been confirmed in the blockchain multiple times.
  • Once funds are in your account, checkout is familiar and straightforward.
Escrow
  • Escrow is provided by the marketplace operator and it is paramount to their business model.
  • Sometimes you’ll find two service tiers: standard escrow (admins are the judges) or multisig escrow.
  • To finance this service (and make a profit), marketplaces charge a small fee.
  • Some sellers are very well established and have stellar reputations. This affords them the privilege of skipping escrow.
  • Oftentimes, sellers will give you a discount (up to 20%) if you skip escrow or finalize early for the benefit of getting paid upfront.
  • You have some number of days (15-30) to finalize the order (at which point the funds are transferred to the seller) or dispute it, at which point the staff gets involved.
Shipping
  • This was not really relevant for my purposes, so I’m not entirely sure how shipping works for drugs. But I did some reading and wanted to share the most creative (emphasis on creative) method for anonymously receiving a package:
  • One user put down the address of his local post office as a shipping address instead of his home. As a recipient, instead of his name he submitted “Holder of Federal Reserve Note number #NNNNN”, #NNNNN being the serial number of a dollar bill in his possession. Apparently he went to the post office holding the bill, correctly identifying himself as the holder of that federal reserve note, and was given the package (which I can only assume contained drugs).
Network effects
  • There are no data network effects in the platform. In fact, deep web marketplace operators want to hold on to as little data as possible, as the opposite increases their exposure to prosecution.
  • The network effects are in the seller’s reputation across many different forums, marketplaces, and websites (including “clear” web services like Reddit).
  • Brand and product drive defensibility. Because the popular sellers are present in all major marketplaces, users mostly make decisions based on product. When new users ask for recommendations, they are oftentimes sent to a particular marketplace because of its ease of use.
Lessons learned
  • A seller’s brand and reputation are extremely important in a system where the intermediary (the marketplace) does not guarantee trust and safety.
  • This is largely decentralized in deep web marketplaces, as vendors make sure their brand is spread across multiple websites and forums.
  • Marketplaces come and go (or get seized by the FBI) but sellers need maintain their reputation.
  • Marketplaces can extract value where they incur costs. Because Bitcoin transactions are a commodity, high take rates and complex fee structures are unsustainable business models. This leads marketplaces to become very thin layers between supply and demand, which commands much smaller transaction fees – as low as 2% – to finance the small set of crucial services (enforcing contracts).
  • The network regulates itself with relatively little involvement from its administrator (if networks are like governments, this is similar to a very small libertarian one).
  • Peer to peer commerce, with no intermediary, can work: it depends on the reputation of the supplier and the size of the discount.
There’s a lot to learn from these platforms as we continue to think about how the Blockchain and other new technologies might impact traditional business models. For example, marketplaces with cost structures that command high take rates are vulnerable to Bitcoin-driven business models with very low or non-existent transaction fees. It could be that what drives adoption of unbundled services is competition by lowering costs.
I’m also wondering how applications could build network effects while defaulting to decentralized open data through the Blockchain Application Stack. While deep web marketplaces don’t fit this model, periodically purging the database has similar implications to giving up control of your user’s information by using decentralized data stores. Perhaps the answer is to have the best product and user experience.
This brings about a very interesting set of questions for both entrepreneurs and investors.
How do you monetize a decentralized network? Is it SAAS on top of the network?
How can you build build network effects while relinquishing control of the data? Do you compete on product and user experience? Is that defensible?
We have some ideas, but no definitive answers.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

How the Deep Web Works Part 7

Lots More Information

Author's Note: How the Deep Web WorksThe Deep Web is a vague, ambiguous place. But while researching this story, it was easy to conclude at least one thing for sure -- most news headlines tend to sensationalize the dark Web and its seedier side, and rarely mention the untapped potential of the deep Web. Articles about illegal drugs and weapons obviously draw more readers than those detailing the technical challenges of harvesting data from the deep Web. Read the negative, breathless articles with a grain of salt. It's worth remembering that there's a whole lot more to the deep Web than the obvious criminal element. As engineers find better, faster ways to catalog the Web's stores of data, the Internet as a whole could transform our society in amazing ways.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

How the Deep Web Works Part 6

The Brighter Side of Darkness

The dark Web has its ominous overtones. But not everything on the dark side is bad. There are all sorts of services that don't necessarily run afoul of the law.
The dark Web is home to alternate search engines, e-mail services, file storage, file sharing, social media, chat sites, news outlets and whistleblowing sites, as well as sites that provide a safer meeting ground for political dissidents and anyone else who may find themselves on the fringes of society.
In an age where NSA-type surveillance is omnipresent and privacy seems like a thing of the past, the dark Web offers some relief to people who prize their anonymity. Dark Web search engines may not offer up personalized search results, but they don't track your online behavior or offer up an endless stream of advertisements, either. Bitcoin may not be entirely stable, but it offers privacy, which is something your credit card company most certainly does not.
For citizens living in countries with violent or oppressive leaders, the dark Web offers a more secure way to communicate with like-minded individuals. Unlike Facebook or Twitter, which are easy for determined authorities to monitor, the dark Web provides deeper cover and a degree of safety for those who would badmouth or plot to undermine politicians or corporate overlords.
A paper written by researchers at the University of Luxembourg attempted to rank the most commonly accessed materials on the dark Web. What they found was that although sites trading in illegal activities and adult content are very popular, so too are those concerned with human rights and freedom of information [Source: ArXiv].
So although the dark Web definitely has its ugly side, it has great potential, too.


Even Deeper

The deep Web is only getting deeper. Its store of human knowledge and trivialities grows more massive every day, complicating our efforts to make sense of it all. In the end, that's perhaps the biggest challenge behind the Internet that we've created.
Programmers will continue to improve search engine algorithms, making them better at delving into deeper layers of the Web. In doing so, they'll help researchers and businesses connect and cross-reference information in ways that were never possible before.
At the same time, the primary job of a smart search engine is not to simply find information. What you really want it to do is find the most relevant information. Otherwise, you're left awash in a sea of cluttered data that leaves you wishing you had never clicked on that search button.
That's the problem of so-called big data. Big data is the name for sets of data that are so large that they become unmanageable and incoherent. Because the Internet is growing so quickly, our whole world is overrun with data, and it's hard for anyone to make sense of it all -- even all of those powerful, all-knowing computers at Bing and Google headquarters.
As the Internet grows, every large company spends more and more money on data management and analysis, both to keep their own organizations functioning and also to obtain competitive advantages over others. Mining and organizing the deep Web is a vital part of those strategies. Those companies that learn to leverage this data for their own uses will survive and perhaps change the world with new technologies. Those that rely only on the surface Web won't be able to compete.
In the meantime, the deep Web will continue to perplex and fascinate everyone who uses the Internet. It contains an enthralling amount of knowledge that could help us evolve technologically and as a species when connected to other bits of information. And of course, its darker side will always be lurking, too, just as it always does in human nature. The deep Web speaks to the fathomless, scattered potential of not only the Internet, but the human race, too.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

How the Deep Web Works Part 5

Titillating Tor

The most infamous of these onion sites was the now-defunct Silk Road, an online marketplace where users could buy drugs, guns and all sorts of other illegal items. The FBI eventually captured Ross Ulbricht, who operated Silk Road, but copycat sites like Black Market Reloaded are still readily available.
Oddly enough, Tor is the result of research done by the U.S. Naval Research Laboratory, which created Tor for political dissidents and whistleblowers, allowing them to communicate without fear of reprisal.
Tor was so effective in providing anonymity for these groups that it didn't take long for the criminally-minded to start using it as well.
That leaves U.S. law enforcement in the ironic position of attempting to track criminals who are using government-sponsored software to hide their trails. Tor, it would seem, is a double-edged sword.
Anonymity is part and parcel on the dark Web, but you may wonder how any money-related transactions can happen when sellers and buyers can't identify each other. That's where Bitcoin comes in on the Deep Web.
If you haven't heard of Bitcoin, it's basically an encrypted digital currency. You can read all about it on How Bitcoin Works. Like regular cash, Bitcoin is good for transactions of all kinds, and notably, it also allows for anonymity; no one can trace a purchase, illegal or otherwise.
Bitcoin may be the currency of the future -- a decentralized and unregulated type of money free of the reins of any one government. But because Bitcoin isn't backed by any government, its value fluctuates, often wildly. It's anything but a safe place to store your life savings. But when paired properly with Tor, it's perhaps the closest thing to a foolproof way to buy and sell on the Web.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

How the Deep Web Works Part 4

Darkness Falls


The deep Web may be a shadow land of untapped potential, but with a bit of skill and some luck, you can illuminate a lot of valuable information that many people worked to archive. On the dark Web, where people purposely hide information, they'd prefer it if you left the lights off.
The dark Web is a bit like the Web's id. It's private. It's anonymous. It's powerful. It unleashes human nature in all its forms, both good and bad.
The bad stuff, as always, gets most of the headlines. You can find illegal goods and activities of all kinds through the dark Web. That includes illicit drugs, child pornography, stolen credit card numbers, human trafficking, weapons, exotic animals, copyrighted media and anything else you can think of. Theoretically, you could even, say, hire a hit man to kill someone you don't like.
But you won't find this information with a Google search. These kinds of Web sites require you to use special software, such as The Onion Router, more commonly known as Tor.
Tor is software that installs into your browser and sets up the specific connections you need to access dark Web sites. Critically, Tor is an encrypted technology that helps people maintain anonymity online. It does this in part by routing connections through servers around the world, making them much harder to track.
Tor also lets people access so-called hidden services -- underground Web sites for which the dark Web is notorious. Instead of seeing domains that end in .com or .org, these hidden sites end in .onion. On the next page we'll peel back the layers of some of those onions.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

How the Deep Web Works Part 3

Deep Potential

Data in the Deep Web is hard for search engines to see, but unseen doesn't equal unimportant. As you can see just from our newspaper example, there's immense value in the information tucked away in the deep Web.
The deep Web is an endless repository for a mind-reeling amount of information. There are engineering databases, financial information of all kinds, medical papers, pictures, illustrations ... the list goes on, basically, forever.
And the deep Web is only getting deeper and more complicated. For search engines to increase their usefulness, their programmers must figure out how to dive into the deep Web and bring data to the surface. Somehow they must not only find valid information, but they must find a way to present it without overwhelming the end users.
As with all things business, the search engines are dealing with weightier concerns than whether you and I are able to find the best apple crisp recipe in the world. They want to help corporate powers find and use the deep Web in novel and valuable ways.
For example, construction engineers could potentially search research papers at multiple universities in order to find the latest and greatest in bridge-building materials. Doctors could swiftly locate the latest research on a specific disease.
The potential is unlimited. The technical challenges are daunting. That's the draw of the deep Web. Yet there's a murkier side to the deep Web, too -- one that's troubling to a lot of people for a lot reasons.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

How the Deep Web Works Part 2

As we've already noted, there are millions upon millions of sub-pages strewn throughout millions of domains. There are internal pages with no external links, such as internal.howstuffworks.com, which are used for site maintenance purposes. There are unpublished or unlisted blog posts, picture galleries, file directories, and untold amounts of content that search engines just can't see.
Here's just one example. There are many independent newspaper Web sites online, and sometimes, search engines index a few of the articles on those sites. That's particularly true for major news stories that receive a lot of media attention. A quick Google search will undoubtedly unveil many dozens of articles on, for example, World Cup soccer teams.
But if you're looking for a more obscure story, you may have to go directly to a specific newspaper site and then browse or search content to find what you're looking for. This is especially true as a news story ages. The older the story, the more likely it's stored only on the newspaper's archive, which isn't visible on the surface Web. Subsequently, that story may not appear readily in search engines -- so it counts as part of the deep Web.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

How the Deep Web Works Part 1

The deep Web is enormous in comparison to the surface Web. Today's Web has more than 555 million registered domains. Each of those domains can have dozens, hundreds or even thousands of sub-pages, many of which aren't cataloged, and thus fall into the category of deep Web.
Although nobody really knows for sure, the deep Web may be 400 to 500 times bigger that the surface Web [source: BrightPlanet]. And both the surface and deep Web grow bigger and bigger every day.
To understand why so much information is out of sight of search engines, it helps to have a bit of background on searching technologies. You can read all about it with How Internet Search Engines Work, but we'll give you a quick rundown here.
Search engines generally create an index of data by finding information that's stored on Web sites and other online resources. This process means using automated spiders or crawlers, which locate domains and then follow hyperlinks to other domains, like an arachnid following the silky tendrils of a web, in a sense creating a sprawling map of the Web.
This index or map is your key to finding specific data that's relevant to your needs. Each time you enter a keyword search, results appear almost instantly thanks to that index. Without it, the search engine would literally have to start searching billions of pages from scratch every time someone wanted information, a process that would be both unwieldy and exasperating.
But search engines can't see data stored to the deep Web. There are data incompatibilities and technical hurdles that complicate indexing efforts. There are private Web sites that require login passwords before you can access the contents. Crawlers can't penetrate data that requires keyword searches on a single, specific Web site. There are timed-access sites that no longer allow public views once a certain time limit has passed.
All of those challenges, and a whole lot of others, make data much harder for search engines to find and index. Keep reading to see more about what separates the surface and deep Web.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

Deep Web Info

What a tangled web we weave, indeed. About 40 percent of the world's population uses the Web for news, entertainment, communication and myriad other purposes [source: Internet World Stats]. Yet even as more and more people log on, they are actually finding less of the data that's stored online. That's because only a sliver of what we know as the World Wide Web is easily accessible.
The so-called surface Web, which all of us use routinely, consists of data that search engines can find and then offer up in response to your queries. But in the same way that only the tip of an iceberg is visible to observers, a traditional search engine sees only a small amount of the information that's available -- a measly 0.03 percent [source: OEDB].
As for the rest of it? Well, a lot of it's buried in what's called the deep Web. The deep Web (also known as the undernet, invisible Web and hidden Web, among other monikers) consists of data that you won't locate with a simple Google search.
No one really knows how big the deep Web really is, but it's hundreds (or perhaps even thousands) of times bigger that the surface Web. This data isn't necessarily hidden on purpose. It's just hard for current search engine technology to find and make sense of it.
There's a flip side of the deep Web that's a lot murkier -- and, sometimes, darker -- which is why it's also known as the dark Web. In the dark Web, users really do intentionally bury data. Often, these parts of the Web are accessible only if you use special browser software that helps to peel away the onion-like layers of the dark Web.
This software maintains the privacy of both the source and the destination of data and the people who access it. For political dissidents and criminals alike, this kind of anonymity shows the immense power of the dark Web, enabling transfers of information, goods and services, legally or illegally, to the chagrin of the powers-that-be all over the world.
Just as a search engine is simply scratching the surface of the Web, we're only getting started. Keep reading to find out how tangled our Web really becomes.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,  

Sunday, January 18, 2015

Guide To Accessing The Deep Web

Guide To Accessing The Deep Web

If you binge-watched the second season of House of Cards, along with a reported 15% of Netflix's 44 million subscribers, you may be newly interested in the Deep Web. Slate has done a good job of describing what the Deep Web is and isn't, but they don't tell you how to get there.

How To Access The Deep Web

First: the hot sheets. Subreddit forums for DeepWeb, onions, and Tor are the way to go in terms of gathering a backgrounder for entry points into DarkNet. Unsurprisingly though, much of the information currently on the surface Internet about the actual underbelly of the web is outdated. Ever since Silk Road's takedown last year, the Under-web has been changing.

To get into the Deep Web these days, you first have to download the Tor add-on for Firefox. By downloading the Tor Browser Bundle from the Tor Project you are securing your anonymity to browse, which is the main draw for using Tor. Once you have downloaded the browser bundle, Tor builds a circuit of encrypted connections through a randomized relay. In layman's terms that means that your online activity is covered as Tor randomly pings your IP address from one place to the other, making whatever you do less traceable.

You might also like: The Casual Bitcoin Buyer's Guide To Investing In Cryptocurrencies

Multiple Redditors urge reading the Tor Project's warning page, where they discourage torrent file sharing and downloading while using Tor. The idea is to follow protocol maintaining your anonymity while browsing, chatting, or navigating. This obviously includes giving away your personal information like email addresses, phone numbers, names, time zones, or home addresses in any context.

The newest iteration of the Tor browser, Tor 4.0, was released in October 2014. It contains a variety of product tweaks designed primarily to enable use behind China's massive Internet firewall.

Other precautions include placing duct tape on your webcam, enabling your computer's firewall, and turning off cookies and JavaScript. Again, here is where you want to be completely free of an identity, so treading cautiously is key. The NSA and other government outlets peruse the Dark Web and onion sites frequently using cross-reference tools, malware, and remote administration tools to de-anonymize users engaging in illegal activity.

While the Deep Web houses the retail of weapons, drugs, and illicit erotica, there are also useful tools for journalists, researchers, or thrill seekers. It's also worth noting that mere access through Tor is not illegal but can arouse suspicion with the law. Illegal transactions usually begin on the Deep Web but those transactions quite often head elsewhere for retail, private dialoguing, or in-person meetups; that's how most people get caught by law enforcement officials.

For mobile users, several browsers exist which purport to—more or less—allow Tor to be used on an Android or iOS device. These browsers include OrWeb, Anonymous Browser Connect Tor, the mobile Firefox add-on, Onion Browser, and Red Onion. However, it is important to note that security concerns have been raised for all of these browsers and that anonymous browsing cannot be 100% guaranteed for any of these.

Where To Go Once You're On The Inside

After reading up on the material, downloading Tor, and logging out of every other application, you can finally open Tor's Browser Bundle to begin secure navigation. Network navigation is slow once you are inside because of the running relay, so expect pages to load at a snail's pace.

The most common suggestion on Reddit is to start at the "Hidden Wiki." The Hidden Wiki has a similar interface as Wikipedia and lists by category different sites to access depending on your interest. Categories include: Introduction Points, News/History, Commercial Services, Forums/Boards/Chans, and H/P/A/W/V/C (Hack, Phreak, Anarchy, Warez, Virus, Crack) just to name a few. Under each of these headings are multiple sites with an onion address and a brief description of what you will find there.

Many of the listed sites on the Hidden Wiki though have been taken down. Deep Web Tor, Tor Jump, Tor Answers, and Tor.info were all busts. When the feds took down Silk Road, many other sites also fell victim and/or are currently down for maintenance. Still, gun, drug, and child porn marketplaces operate even though they are on much smaller scales and with a fraction of the reach than that of Silk Road or Atlantis, another drug-peddling site.

Some pages are less nefarious, but arouse your curiosity nonetheless. StaTors.Net is the Twitter for Tor users and Hell Online is the antisocial network with 369 members and 15 different groups. Torchan resembles Reddit, though you need to enter the username and password torchan2 for access, and is still up and running. But recent activity except in Request and Random rooms has all but stopped.

In the Random room a user asked for a new link to Silk Road and the responses were limited. Another user posting an image of a child fully clothed featuring bare feet pleading for a site featuring underage bare feet. An Anonymous user responded: "Someone please give this guy a link, this poor guy has been looking/asking for over a month now."

One popular chat service is OnionChat, an anonymous Tor-based real-time chat room quasi-affiliated with the Onions subreddit. The project's code is available on GitHub as well.

The DeepWeb Link Directory in the site OnionDir had some promising hyperlinks and some not-so-promising ones like the now defunct Deep Web Radio and a blog claiming to be a Deep Web blog but was actually just stories dedicated to spanking.

The New Yorker Strongbox is a secure transmission for writers and editors where I was given the code name: riddle yeah abreacts murgeoning. Through a given codename you can submit a message and/or file to the New Yorker's editorial staff. Mike Tigas, a news application developer for ProPublica, has a functioning blog in the Deep Web but has not posted anything new for some time, which was true for many other blogs as well.

New Sites To Explore

Reddit user NekroTor is on a quest to reboot many of the Freedom Hosting sites that were taken down. On February 16th of this year, on his onion-routed blog, Nekrotown, he wrote, "2 days ago the BlackMarket Reloaded forum got seized. On the same day, the long-awaited Utopia Market was seized, which just goes to show that all the markets fucking suck these days except for Agora and TMP, and that you should just wait until BlackMarket Reloaded opens up again... eventually ...5 years later, no BMR."

NekroTor is correct in writing that most of the content right now on the Dark Web is not that great. On top of the fact that there used to be a wealth of sites for illegal black market interactions, there also used to be radio, books, blogs, political conversations, and even an Encyclopedia Dramatica that was a satirical culture-based wiki and is now laden with porn and pop-ups.

NekroTor created a new version of Hidden Wiki that has some functional links to audio and video streaming as well as some up-to-date forums for socializing and buying and selling. There are still a few image boards left, but the popular Onii-chan has the words "Well be back later" typed over spinning dildos.

Redditors Who Are Reaching Out For Deep Web Direction

After watching House of Cards, user TrelianScar turned to Reddit for guidance on how to navigate the Dark Web. TrelianScar is not alone. The Deep Web is making appearances in the media, in dinner conversations, and of course on Internet forums. One user jokingly writes to TrelianScar saying, "Wait till we send you an iPad. Then talk to the Dutch oil painting. Then await instructions," referencing HOC's unrealistic depiction of Deep Web interactions.

On a more serious note though, user Serbia_Strong writes, "What are you looking for first of all? Drugs? Guns? Assassins? Credit cards or counterfeit cash? I'd start your journey at the Hidden Wiki and then narrow in on your interests. I pretty much save every site I come across (you can't exactly just google them). Start at The Hidden Wiki and if you need any links just ask. Enjoy your descent into madness :)"

Another user, Dexter-Del-Rey explained a similar conundrum last week—he too is new to the Deep Web and wants some functional starter links. Redditor Ampernand writes back saying, "On the topic of torchan... here's a good piece on how it fell authored by the previous host. Currently torchan is hosted by someone that allows cp, gore etc, censors critics and doesn't give a flying fuck about the community. Effectively torchan has become exactly what it was trying to not be. Also, nntpchan is better." Ampernand links to NNTP-chan, which is a new forum replacing the image board Onii-chan.

New channels are popping up daily in the Deep Web. Currently, marketplace alternatives to Silk Road, Agora, and Pandora are the most frequented. Nonetheless, both TrelianScar and Dexter-Del-Rey were each respectively warned in their threads that the Dark Web is chock-full of scammers and is quite unlike its Hollywood depiction.

Interestingly, the Deep Web has lost much of its stigma over the past year and mainstream web services are experimenting with the platform even if they're not embracing it. In late October 2014, Facebook enabled Tor browser users to visit them anonymously, saying in a press release that "It’s important to us at Facebook to provide methods for people to use our site securely."

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search, 

Deep Web

Deep Web (also called the Deepnet,[1] Invisible Web,[2] or Hidden Web[3]) is the portion of World Wide Web content that is not indexed by standard search engines. Mike Bergman, founder of BrightPlanet and credited with coining the phrase,[4] said that searching on the Internet today can be compared to dragging a net across the surface of the ocean: a great deal may be caught in the net, but there is a wealth of information that is deep and therefore missed.[5] Most of the Web's information is buried far down on sites, and standard search engines do not find it. Traditional search engines cannot see or retrieve content in the deep Web. The portion of the Web that is indexed by standard search engines is known as the Surface Web. As of 2001,[needs update] the deep Web was several orders of magnitude larger than the surface Web.[6] The deep web should not be confused with the dark Internet, computers that can no longer be reached via the Internet. The Darknet distributed file sharing network, can be classified as part of the Deep Web. Although much of the Deep Web is innocuous, some prosecutors and government agencies, among others, are concerned that the Deep Web is a haven for serious criminality.[7] Contents 1 Size 2 Naming 3 Methods 4 Indexing the Deep Web 5 Classifying resources Size Bright Planet, a web-services company, describes the size of the Deep Web in this way: It is impossible to measure or put estimates onto the size of the deep web because the majority of the information is hidden or locked inside databases. Early estimates suggested that the deep web is 400 to 550 times larger than the surface web. However, since more information and sites are always being added, it can be assumed that the deep web is growing exponentially at a rate that cannot be quantified. Estimates based on extrapolations from a study done at University of California, Berkeley in 2001[6] speculate that the deep web consists of about 7.5 petabytes. More accurate estimates are available for the number of resources in the deep Web: research of He et al. detected around 300,000 deep web sites in the entire Web in 2004,[8] and, according to Shestakov, around 14,000 deep web sites existed in the Russian part of the Web in 2006.[9] Naming Bergman, in a seminal paper on the deep Web published in The Journal of Electronic Publishing, mentioned that Jill Ellsworth used the term invisible Web in 1994 to refer to websites that were not registered with any search engine.[6] Bergman cited a January 1996 article by Frank Garcia:[10] It would be a site that's possibly reasonably designed, but they didn't bother to register it with any of the search engines. So, no one can find them! You're hidden. I call that the invisible Web. Another early use of the term Invisible Web was by Bruce Mount and Matthew B. Koll of Personal Library Software, in a description of the @1 deep Web tool found in a December 1996 press release.[11] The first use of the specific term Deep Web, now generally accepted, occurred in the aforementioned 2001 Bergman study.[6] Methods Methods which prevent web pages from being indexed by traditional search engines may be categorized as one or more of the following: Dynamic content: dynamic pages which are returned in response to a submitted query or accessed only through a form, especially if open-domain input elements (such as text fields) are used; such fields are hard to navigate without domain knowledge. Unlinked content: pages which are not linked to by other pages, which may prevent Web crawling programs from accessing the content. This content is referred to as pages without backlinks (also known as inlinks). Also, search engines do not always detect all backlinks from searched web pages. Private Web: sites that require registration and login (password-protected resources). Contextual Web: pages with content varying for different access contexts (e.g., ranges of client IP addresses or previous navigation sequence). Limited access content: sites that limit access to their pages in a technical way (e.g., using the Robots Exclusion Standard or CAPTCHAs, or no-store directive which prohibit search engines from browsing them and creating cached copies.[12]) Scripted content: pages that are only accessible through links produced by JavaScript as well as content dynamically downloaded from Web servers via Flash or Ajax solutions. Non-HTML/text content: textual content encoded in multimedia (image or video) files or specific file formats not handled by search engines. Software: Certain content is intentionally hidden from the regular internet, accessible only with special software, such as Tor. Tor allows users to access websites using the .onion host suffix anonymously, hiding their IP address. Other such software includes I2P and darknet software. Indexing the Deep Web While it is not always possible to directly discover a specific web server's content so that it may be indexed, a site potentially can be accessed indirectly (due to computer vulnerabilities). To discover content on the Web, search engines use web crawlers that follow hyperlinks through known protocol virtual port numbers. This technique is ideal for discovering content on the surface Web but is often ineffective at finding Deep Web content. For example, these crawlers do not attempt to find dynamic pages that are the result of database queries due to the indeterminate number of queries that are possible.[4] It has been noted that this can be (partially) overcome by providing links to query results, but this could unintentionally inflate the popularity for a member of the deep Web. DeepPeep, Intute, Deep Web Technologies, Scirus, and Ahmia.fi are a few search engines that have accessed the Deep Web. Intute ran out of funding and is now a temporary static archive as of July, 2011.[13] Scirus retired near the end of January, 2013.[14] Researchers have been exploring how the Deep Web can be crawled in an automatic fashion, including content that can be accessed only by special software such as Tor. In 2001, Sriram Raghavan and Hector Garcia-Molina (Stanford Computer Science Department, Stanford University)[15][16] presented an architectural model for a hidden-Web crawler that used key terms provided by users or collected from the query interfaces to query a Web form and crawl the Deep Web content. Alexandros Ntoulas, Petros Zerfos, and Junghoo Cho of UCLA created a hidden-Web crawler that automatically generated meaningful queries to issue against search forms.[17] Several form query languages (e.g., DEQUEL[18]) have been proposed that, besides issuing a query, also allow extraction of structured data from result pages. Another effort is DeepPeep, a project of the University of Utah sponsored by the National Science Foundation, which gathered hidden-Web sources (Web forms) in different domains based on novel focused crawler techniques.[19][20] Commercial search engines have begun exploring alternative methods to crawl the deep Web. The Sitemap Protocol (first developed, and introduced by Google in 2005) and mod oai are mechanisms that allow search engines and other interested parties to discover deep Web resources on particular Web servers. Both mechanisms allow Web servers to advertise the URLs that are accessible on them, thereby allowing automatic discovery of resources that are not directly linked to the surface Web. Google's deep Web surfacing system pre-computes submissions for each HTML form and adds the resulting HTML pages into the Google search engine index. The surfaced results account for a thousand queries per second to deep Web content.[21] In this system, the pre-computation of submissions is done using three algorithms: selecting input values for text search inputs that accept keywords, identifying inputs which accept only values of a specific type (e.g., date), and selecting a small number of input combinations that generate URLs suitable for inclusion into the Web search index. In 2008, to facilitate users of Tor hidden services in their access and search of a hidden .onion suffix, Aaron Swartz designed Tor2web—a proxy application able to provide access by means of common web browsers.[22] Using this application, Deep Web links appear as a random string of letters followed by the .onion TLD. For example, http://xmh57jrzrnw6insl followed by .onion, links to TORCH, the Tor search engine web page. Classifying resources Most of the work of classifying search results has been in categorizing the surface Web by topic. For classification of deep Web resources, Ipeirotis et al.[23] presented an algorithm that classifies a deep Web site into the category that generates the largest number of hits for some carefully selected, topically-focused queries. Deep Web directories under development include OAIster at the University of Michigan, Intute at the University of Manchester, Infomine[24] at the University of California at Riverside, and DirectSearch (by Gary Price). This classification poses a challenge while searching the deep Web whereby two levels of categorization are required. The first level is to categorize sites into vertical topics (e.g., health, travel, automobiles) and sub-topics according to the nature of the content underlying their databases. The more difficult challenge is to categorize and map the information extracted from multiple deep Web sources according to end-user needs. Deep Web search reports cannot display URLs like traditional search reports. End users expect their search tools to not only find what they are looking for, but to be intuitive and user-friendly. In order to be meaningful, the search reports have to offer some depth to the nature of content that underlie the sources or else the end-user will be lost in the sea of URLs that do not indicate what content lies beneath them. The format in which search results are to be presented varies widely by the particular topic of the search and the type of content being exposed. The challenge is to find and map similar data elements from multiple disparate sources so that search results may be exposed in a unified format on the search report irrespective of their source.

how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,

100 Tips and Tools to Search the Deep Web

 100 Tips and Tools to Search the Deep Web

Experts say that typical search engines like Yahoo! and Google only pick up about 1% of the information available on the Internet. The rest of that information is considered to be hidden in the deep web, also referred to as the invisible web. So how can you find all the rest of this information? This list offers 100 tips and tools to help you get the most out of your Internet searches.
Meta-Search Engines
Meta-search engines use the resources of many different search engines to gather the most results possible. Many of these will also eliminate duplicates and classify results to enhance your search experience.
  1. SurfWax. This search engine works very well for reaching deep into the web for information.
  2. Academic Index. Created by the former chair of Texas Association of School Librarians, this meta-search engine only pulls from databases and resources that are approved by librarians and educators.
  3. Clusty. Clusty searches through top search engines, then clusters the results so that information that may have been hidden deep in the search results is now readily available.
  4. Dogpile. Dogpile searches rely on several top search engines for the results then removes duplicates and strives to present only relevant results.
  5. Turbo 10. This meta-search engine is specifically designed to search the deep web for information.
  6. Multiple Search. Save yourself the work by using this search engine that looks among major search engines, social networks, flickr, Wikipedia, and many more sites.
  7. Mamma. Click on the Power Search option to customize your search experience with this meta-search engine.
  8. World Curry Guide. This meta-search tool with a strong European influence has been around since 1997 and is still growing strong.
  9. Fazzle.com. Give this meta-search engine a try. It accesses a large number of databases and claims to have more access to information than Google.
  10. Icerocket. Search blogs as well as the general Internet, MySpace, the news, and more to receive results by posting date.
  11. iZito. Get results from a variety of major search engines that come to you clustered in groups. You can also receive only US website results or receive results with a more international perspective.
  12. Ujiko. This unusual meta-search tool allows for you to customize your searches by eliminating results or tagging some as favorites.
Semantic Search Tools and Databases
Semantic search tools depend on replicating the way the human brain thinks and categorizes information to ensure more relevant searches. Give some of these semantic tools and databases a try.
  1. Hakia. This popular semantic search engine only accesses websites that are recommended by librarians.
  2. Zotero. Firefox users will like this add-on that helps you organize your research material by collecting, managing, and citing any references from Internet research.
  3. Freebase. This community-powered database includes information on millions of topics.
  4. Powerset. Enter a topic, phrase, or question to find information from Wikipedia with this semantic application.
  5. Kartoo. Enter any keyword to receive a visual map of the topics that pertain to your keyword. Hover your mouse over each to get a thumbnail of the website.
  6. DBpedia. Another Wikipedia resource, ask complex questions with this semantic program to get results from within Wikipedia.
  7. Quintura. Entering your search term will create a cloud of related terms as well as a list of links. Hover over one of the words or phrases in the cloud to get an entirely different list of links.
  8. [true knowledge]. Help with current beta testing at this search engine or try their Quiz Bot that finds answers to your questions.
  9. Stumpedia. This search engine relies on its users to index, organize, and review information coming from the Internet.
  10. Evri. This search engine provides you with highly relevant results from articles, papers, blogs, images, audio, and video on the Internet.
  11. Gnod. When you search for books, music, movies and people on this search engine, it remembers your interests and focuses the search results in that direction.
  12. Boxxet. Search for what interests you and you will get results from the "best of" news, blogs, videos, photos, and more. Type in your keyword and in addition to the latest news on the topic, you will also receive search results, online collections, and more.
General Search Engines and Databases
These databases and search engines for databases will provide information from places on the Internet most typical search engines cannot.
  1. DeepDyve. One of the newest search engines specifically targeted at exploring the deep web, this one is available after you sign up for a free membership.
  2. OAIster. Search for digital items with this tool that provides 12 million resources from over 800 repositories.
  3. direct search. Search through all the direct search databases or select a specific one with this tool.
  4. CloserLook Search. Search for information on health, drugs and medicine, city guides, company profiles, and Canadian airfares with this customized search engine that specializes in the deep web.
  5. Northern Light Search. Find information with the quick search or browse through other search tools here.
  6. Yahoo! Search Subscriptions. Use this tool to combine a search on Yahoo! with searches in journals where you have subscriptions such as Wall Street Journal and New England Journal of Medicine.
  7. CompletePlanet. With over 70,000 databases and search engines at its disposal, this is an excellent resource for searching the deep web.
  8. The Scout Archives. This database is the culmination of nine years’ worth of compiling the best of the Internet.
  9. Daylife. Find news with this site that offers some of the best global news stories along with photos, articles, quotes, and more.
  10. Silobreaker. This tool shows how news and people in the news impacts the global culture with current news stories, corresponding maps, graphs of trends, networks of related people or topics, fact sheets, and more.
  11. spock. Find anyone on the web who might not normally show up on the surface web through blogs, pictures, social networks, and websites here.
  12. The WWW Virtual Library. One of the oldest databases of information available on the web, this site allows you to search by keyword or category.
  13. pipl. Specifically designed for searching the deep web for people, this search engine claims to be the most powerful for finding someone.
Academic Search Engines and Databases
The world of academia has many databases not accessible by Google and Yahoo!, so give these databases and search engines a try if you need scholarly information.
  1. Google Scholar. Find information among academic journals with this tool.
  2. WorldCat. Use this tool to find items in libraries including books, CDs, DVDs, and articles.
  3. getCITED. This database of academic journal articles and book chapters also includes a discussion forum.
  4. Microsoft Libra. If you are searching for computer science academic research, then Libra will help you find what you need.
  5. BASE – Bielefeld Academic Search Engine. This multi-disciplinary search engine focuses on academic research and is available in German, Polish, and Spanish as well as English.
  6. yovisto. This search engine is an academic video search tool that provides lectures and more.
  7. AJOL – African Journals Online. Search academic research published in AJOL with this search engine.
  8. HighWire Press. From Stanford, use this tool to access thousands of peer-reviewed journals and full-text articles.
  9. MetaPress. This tool claims to be the "world’s largest scholarly content host" and provides results from journals, books, reference material, and more.
  10. OpenJ-Gate. Access over 4500 open journals with this tool that allows you to restrict your search to peer-reviewed journals or professional and industry journals.
  11. Directory of Open Access Journals. This journal search tool provides access to over 3700 top "quality controlled" journals.
  12. Intute. The resources here are all hand-selected and specifically for education and research purposes.
  13. Virtual Learning Resource Center. This tool provides links to thousands of academic research sites to help students at any level find the best information for their Internet research projects.
  14. Gateway to 21st Century Skills. This resource for educators is sponsored by the US Department of Education and provides information from a variety of places on the Internet.
  15. MagBot. This search engine provides journal and magazine articles on topics relevant to students and their teachers.
  16. Michigan eLibrary. Find full-text articles as well as specialized databases available for searching.
Scientific Search Engines and Databases
The scientific community keeps many databases that can provide a huge amount of information but may not show up in searches through an ordinary search engine. Check these out to see if you can find what you need to know.
  1. Science.gov. This search engine offers specific categories including agriculture and food, biology and nature, Earth and ocean sciences, health and medicine, and more.
  2. WorldWideScience.org. Search for science information with this connection to international science databases and portals.
  3. CiteSeer.IST. This search engine and digital library will help you find information within scientific literature.
  4. Scirus. This science search engine moves beyond journal articles and also includes searches among such resources as scientists’ webpages, courseware, patents, and more.
  5. Scopus. Find academic information among science, technology, medicine, and social science categories.
  6. GoPubMed. Search for biomedical texts with this search engine that accesses PubMed articles.
  7. the Gene Ontology. Search the Gene Ontology database for genes, proteins, or Gene Ontology terms.
  8. PubFocus. This search engine searches Medline and PubMed for information on articles, authors, and publishing trends.
  9. Scitopia. This "deep federated search" brings the best information from the fields of science and technology.
  10. Scitation. Find over one million scientific papers from journals, conferences, magazines, and other sources with this tool.
Custom Search Engines
Custom search engines narrow your focus and eliminate quite a bit of the extra information usually contained in search results. Use these resources to find custom search engines or use the specific custom search engines listed below.
  1. CustomSearchEngine.com. This listing includes many of the Google custom search engines created.
  2. CustomSearchGuide.com. Find custom search engines here or create your own.
  3. CSE Links. Use this site to find Google Coop custom search engines.
  4. PGIS PPGIS Custom Search. This search engine is customized for those interested in the "practice and science" of PGIS/PPGIS.
  5. Files Tube. Search for files in file sharing and uploading sites with this search engine.
  6. Trailmonkey’s Custom Search Engine. This outdoor adventure search engine will help find information such as trails, maps, and wildlife around the world.
  7. Rollyo. "Roll your own search engine" at this site where you determine which sites will be included in your searches.
  8. Webhoker.com. Use this custom search engine to find information about Northern Ireland.
  9. Figure Skating Custom Search Engine. Use this search engine to learn about figure skating. The more this search engine is used, the better the results become.
  10. Custom Search Engines. There are three custom search engines here, two of which may be relevant for anyone interested in Utah constitution or juvenile justice.
  11. Go Pets America Custom Search Engine. This search engine will help you find information on pets and animals, their health and wellness, jobs in the field, and more.
Collaborative Information and Databases
One of the oldest forms of information dissemination is word-of-mouth, and the Internet is no different. With the popularity of bookmarking and other collaborative sites, obscure blogs and websites can gain plenty of attention. Follow these sites to see what others are reading.
  1. Del.icio.us. As readers find interesting articles or blog posts, they can tag, save, and share them so that others can enjoy the content as well.
  2. Digg. As people read blogs or websites, they can "digg" the ones they like, thus creating a network of user-selected sites on the Internet.
  3. Technorati. Not only is this site a blog search engine, but it is also a place for members to vote and share, thus increasing the visibility for blogs.
  4. StumbleUpon. As you read information on the Internet, you can Stumble it and give it a thumbs up or down. The more you Stumble, the more closely aligned to your taste will the content become.
  5. Reddit. Working similarly to StumbleUpon, Reddit asks you to vote on articles, then customizes content based on your preferences.
  6. Twine. With Twine you can search for information as well as share with others and get recommendations from Twine.
  7. Kreeo.com. This collaborative site offers shared knowledge from its members through forums, blogs, and shared websites.
  8. Talk Digger. Find information on the Internet based on what others are saying about it. Members discuss web sites, blogs, and specific topics here.
Tips and Strategies
Searching the deep web should be done a bit differently, so use these strategies to help you get started on your deep web searching.
  1. Don’t rely on old ways of searching. Become aware that approximately 99% of content on the Internet doesn’t show up on typical search engines, so think about other ways of searching.
  2. Search for databases. Using any search engine, enter your keyword alongside "database" to find any searchable databases (for example, "running database" or "woodworking database").
  3. Get a library card. Many public libraries offer access to research databases for users with an active library card.
  4. Stay informed. Reading blogs or other updated guides about Internet searches on a regular basis will ensure you are staying updated with the latest information on Internet searches.
  5. Search government databases. There are many government databases available that have plenty of information you may be seeking.
  6. Bookmark your databases. Once you find helpful databases, don’t forget to bookmark them so you can always come back to them again.
  7. Practice. Just like with other types of research, the more you practice searching the deep web, the better you will become at it.
  8. Don’t give up. Researchers agree that most of the information hidden in the deep web is some of the best quality information available.
Helpful Articles and Resources for Deep Searching
Take advice from the experts and read these articles, blogs, and other resources that can help you understand the deep web.
  1. Deep Web – Wikipedia. Get the basics about the deep web as well as links to some helpful resources with this article.
  2. Deep Web – AI3:::Adaptive Information. This assortment of articles from the co-coiner of the phrase "deep web," Michael Bergman offers a look at the current state of deep web perspectives.
  3. The Invisible Web. This article from About.com provides a very simple explanation of the deep web and offers suggestions for tackling it.
  4. ResourceShelf. Librarians and researchers come together to share their findings on fun, helpful, and sometimes unusual ways to gather information from the web.
  5. Docuticker. This blog offers the latest publications from government agencies, NGOs, think tanks, and other similar organizations. Many of these posts are links to databases and research statistics that may not appear so easily on typical web searches.
  6. TechDeepWeb.com. This site offers tips and tools for IT professionals to find the best deep web resources.
  7. Digital Image Resources on the Deep Web. This article includes links to many digital image resources that probably won’t show up on typical search engine results.
  8. Federated Search 101. Learn about federated search tools in this article that will be helpful to businesses thinking about purchasing a federated search product.
  9. Timeline of events related to the Deep Web. This timeline puts the entire history of the deep web into perspective as well as offers up some helpful links.
  10. The Deep Web. Learn terminology, get tips, and think about the future of the deep web with this article.
 Start searching the Deep Web today! Use Tor!
how to access the deep web, deep web search engines, deep web tor, deep web como entrar, deep web search,