Archive for janeiro \31\UTC 2010|Monthly archive page
O uso que a BBC tem feito da web semântica permite não apenas ao leitor encontrar mais rapidamente informação na interface da rede de tevê inglesa e conteúdos relacionados fora do espaço BBC como transformar a base de dados em APIs para criação de novos conteúdos. Reproduzo aqui o trecho que me parece mais contundente. A íntegra está no w3.org.
“The BBC Search Team is building on this newly-available linked data, by creating pan-BBC aggregations of content. These new pages are called Search+ —an indication that we are using this data to enhance the standard search experience. Each Search+ page shows the best content on its particular topic from around bbc.co.uk, and sometimes selected content from outside the BBC. The set of topics that make up the Search+ Pages essentially form a new BBC Controlled Vocabulary (CV) of concepts and entities. Each term in the BBC CV has an associated DBpedia resource, to enable us to use some of the metadata within DBpedia, and also enable links between our CV terms (and associated Search+ pages) and content both inside and outside the BBC. We also intend to use sources other than just DBpedia (e.g. Musicbrainz and Geonames) to provide these “Linked Open Data” associations in the near future.
Various groups around the BBC have contributed to the Search+ project, and we have a fledgling suite of tools to allow people within the BBC to manage the life cycle of the Search+ pages and the data used in them. These tools allow us to: associate DBpedia resources with content pages; then promote those DBpedia resources into the BBC CV in a controlled way; build the associated Search+ pages for those terms in the BBC CV; include additional pieces of content on those Search+ pages—both from inside and outside the BBC; and to monitor the quality and usage of the Search+ pages.
Creating web identifiers for every item the BBC has an interest in, and considering those as aggregations of BBC content about that item, allows us to enable very rich cross-domain user journeys. This means BBC content can be discovered by users in many different ways, and content teams within the organisation have a focal point around which to organise their content.
The RDF representations of these web identifiers allow developers to use our data to build applications. The two issues, providing cross-domain navigation and machine-readable representations, are tightly interleaved. Giving access to machine-readable representations that hold links to further such representations, crossing domain boundaries, means that much richer applications can be built on top of our data, including new BBC products. In addition the system gives us a flexibility and a maintainability benefit: our web site becomes our API. Considering our feeds as an integral part of building a web site also means that they are very cheap to generate: they are just a different view of our data.
The approach has also proved to be an efficient one—allowing different development teams to concentrate on different domains while at the same time benefiting from the activities of the other teams. The small pieces loosely joined approach, which is manifest in any Linked Data project, significantly reduces the need to coordinate teams while at the same time allowing each team to benefit from the activities of others.
Key Benefits of Using Semantic Web Technology
- Usability—Making a site around the things people care and think about.
- User Experience—Having meaningful predicates and granular, addressable resources, so that those resources can be visualised in new ways.
- User Journeys—Allowing users to make their own journeys across our content. On the BBC /nature, users can start making their own documentaries. They can start on an animal, watch a programme clip, follow a link to a related habitat, read about that habitat and so on…
- One page per thing—Making our resources part of the Web and therefore linkable and discoverable.
- Our web site is our API—One URI for both machines and web browsers. Our web site can be used by third parties to create new products, e.g.,URIPlay, TestTubeTelly, FanHubz or Channelography(*).
- Loosely coupled development—Different teams can work together in a loosely coupled fashion. Each team focuses on their domain of interest.”
90 trillion – The number of emails sent on the Internet in 2009.
247 billion – Average number of email messages per day.
1.4 billion – The number of email users worldwide.
100 million – New email users since the year before.
81% – The percentage of emails that were spam.
92% – Peak spam levels late in the year.
24% – Increase in spam since last year.
200 billion – The number of spam emails per day (assuming 81% are spam).
234 million – The number of websites as of December 2009.
47 million – Added websites in 2009.
13.9% – The growth of Apache websites in 2009.
-22.1% – The growth of IIS websites in 2009.
35.0% – The growth of Google GFE websites in 2009.
384.4% – The growth of Nginx websites in 2009.
-72.4% – The growth of Lighttpd websites in 2009.
81.8 million – .COM domain names at the end of 2009.
12.3 million – .NET domain names at the end of 2009.
7.8 million – .ORG domain names at the end of 2009.
76.3 million – The number of country code top-level domains (e.g. .CN, .UK, .DE, etc.).
187 million – The number of domain names across all top-level domains (October 2009).
8% – The increase in domain names since the year before.
1.73 billion – Internet users worldwide (September 2009).
18% – Increase in Internet users since the previous year.
738,257,230 – Internet users in Asia.
418,029,796 – Internet users in Europe.
252,908,000 – Internet users in North America.
179,031,479 – Internet users in Latin America / Caribbean.
67,371,700 – Internet users in Africa.
57,425,046 – Internet users in the Middle East.
20,970,490 – Internet users in Oceania / Australia.
126 million – The number of blogs on the Internet (as tracked by BlogPulse).
84% – Percent of social network sites with more women than men.
27.3 million – Number of tweets on Twitter per day (November, 2009)
57% – Percentage of Twitter’s user base located in the United States.
4.25 million – People following @aplusk (Ashton Kutcher, Twitter’s most followed user).
350 million – People on Facebook.
50% – Percentage of Facebook users that log in every day.
500,000 – The number of active Facebook applications.
4 billion – Photos hosted by Flickr (October 2009).
2.5 billion – Photos uploaded each month to Facebook.
30 billion – At the current rate, the number of photos uploaded to Facebook per year.
1 billion – The total number of videos YouTube serves in one day.
12.2 billion – Videos viewed per month on YouTube in the US (November 2009).
924 million – Videos viewed per month on Hulu in the US (November 2009).
182 – The number of online videos the average Internet user watches in a month (USA).
82% – Percentage of Internet users that view videos online (USA).
39.4% – YouTube online video market share (USA).
81.9% – Percentage of embedded videos on blogs that are YouTube videos.
148,000 – New zombie computers created per day (used in botnets for sending spam, etc.)
2.6 million – Amount of malicious code threats at the start of 2009 (viruses, trojans, etc.)
921,143 – The number of new malicious code signatures added by Symantec in Q4 2009.
Data sources: Website and web server stats from Netcraft. Domain name stats from Verisign and Webhosting.info. Internet user stats from Internet World Stats. Web browser stats from Net Applications. Email stats from Radicati Group. Spam stats from McAfee. Malware stats from Symantec (and here) and McAfee. Online video stats from Comscore, Sysomos and YouTube. Photo stats from Flickr and Facebook. Social media stats from BlogPulse, Pingdom (here and here), Twittercounter, Facebook and GigaOm.