I’ve been thinking about one particular artifact of the folksonomy phenomenon — the folksonomy menu that serves as a sort of buzz index providing users with a quick visualization of the most popular tags (technically I think it’s called a weighted list). Popular tags are displayed in a larger font and it’s relatively easy to identify hot topics at a glance. This visual representation of the popularity of any given tag is undeniably cool. However, once the coolness factor wears off it becomes fairly obvious that these menus are also not very accessible.
The Atom API is an emerging interface for editing content. The interface is RESTful and uses XML and HTTP to define an editing scheme that's easy to implement and extend. History, basic operation, and applications to areas outside weblogs will be covered.
In the information age it is widely understood that there is now too much information. Some of this newly created information will most certainly be valuable, but despite marked improvement in search tools, finding the valuable information is a slow panhandle. Perhaps in light of this situation, the W3C under the direction of Berners-Lee has begun to build the foundation for the next phase of the web. This phase, called the Semantic Web, will make information stored with this technology much more processible by machines.
In this article I will look at the doctype in a lot more detail, showing what it does and how it helps you validate your HTML, how to choose a doctype for your document, and the XML declaration, which you’ll rarely need, but will sometimes come across.
In this paper, we present a novel method for the classification of Web sites. This method exploits both structure and content of Web sites in order to discern their functionality. It allows for distinguishing between eight of the most relevant functional classes of Web sites. We show that a pre-classification of Web sites utilizing structural properties considerably improves a subsequent textual classification with standard techniques. We evaluate this approach on a dataset comprising more than 16,000 Web sites with about 20 million crawled and 100 million known Web pages. Our approach achieves an accuracy of 92% for the coarse-grained classification of these Web sites.
Once upon a time, we were curious and everything we encountered was new. We were excited about discovering new things and the world offered unlimited possibilities. Then we went to school and were taught to color inside the lines, that everything had its place and the world was ordered.
Little things mean a lot. Especially online. Microcontent—or the headlines, decks, subheads and other 'small' pieces of web copy—actually do most of the communicating on your web site. Handled poorly, microcontent can confuse and frustrate web visitors. Here's how to write microcontent to communicate to—instead of discombobulate—your readers.
The purpose of this 1995 paper is to define what is meant by facet analysis, and to review briefly the history of facet analysis within the context of other types of subject analysis in libraries and within the context of information retrieval research.
The weighted list, known popularly as a `tag cloud', has appeared on many popular folksonomy-based web-sites. Flickr, Delicious, Technorati and many others have all featured a tag cloud at some point in their history. However, it is unclear whether the tag cloud is actually useful as an aid to finding information. We conducted an experiment, giving participants the option of using a tag cloud or a traditional search interface to answer various questions. We found that where the information-seeking task required specific information, participants preferred the search interface. Conversely, where the information-seeking task was more general, participants preferred the tag cloud. While the tag cloud is not without value, it is not sufficient as the sole means of navigation for a folksonomy-based dataset.
One of the critical-mass elements for location-based services to actually be useful is for online content to have a sense of geographic context. We're already seeing it to some extent with services such as Flickr allowing photographs to be tagged with GPS coordinates: camera-phones with built-in GPS can automatically tag each photo with the exact location at which it was taken, and that meta-data can then be used to search for photos of a particular area or place.
This article deals with a part of the HTML document that does not get the attention it deserves: the markup that goes inside the head element. By the end of this tutorial you’ll have learnt about the different parts of this section and what they all do, including the doctype, title element, keywords and description (which are handled by meta elements).
Until recently, the Semantic Web was little more than a name for the next-generation Web infrastructure as envisioned by its inventor, Tim Berners-Lee. With the introduction of XML and RDF, and new developments such as RDF Schema and DAML+OIL, the Semantic Web is rapidly taking shape. This paper gives an overview of the state-of-the-art in Semantic Web technology, the key relationships with traditional hypermedia research, and a comprehensive reference list to various sets of literature (hypertext, Web and Semantic Web). A research agenda describes the open research issues in the development of the Semantic Web from the perspective of hypermedia research.
The web is designed to be consumed by humans, and much of the rich, useful information our websites contain, is inaccessible to machines. People can cope with all sorts of variations in layout, spelling, capitalization, color, position, and so on, and still absorb the intended meaning from the page. Machines, on the other hand, need some help. A new kind of web—a semantic web—would be made up of information marked up in such a way that software can also easily understand it. Before considering how we might achieve such a web, let’s look at what we might be able to do with it.
When it comes to the Web, there is nothing more misunderstood than metadata. Technical people search vainly for a way to automate its creation. Many editors and writers want nothing to do with it. And yet without quality metadata a website cannot properly achieve its objectives. It’s time to get serious about metadata.
The keywords and phrases you use in your Meta description tag don't affect your page's ranking in the search engines (for the most part), but this tag can still come in handy in your overall SEO campaigns.
This article provides an overview of work completed at Tsinghua University Library in which a metadata framework was developed to aid in the preservation of digital resources. The metadata framework is used for the creation of metadata to describe resources, and includes an encoding standard used to store metadata and resource structures in information systems. The author points out that the Tsinghua University Library metadata framework provides a successful digital preservation solution that may be an appropriate solution for other organizations as well.
Metadata is one of the most misunderstood aspects of content management and website design. Editors and writers tend to look at it as a technical issue. Technical people look for a software solution. Both are wrong. Metadata is a fundamental skill that web writers and editors must acquire.
Creating great metadata for your content begins with understanding who your reader is. What is the metadata they look for when they read a page of your content? What are the type of words they use when they search for your content? When scanning your classification, what are the "trigger words" that will make them want to go deeper into your website?
The shift in how search engines treat keywords is significant. They tend to ignore the keyword metatag and rather look for keywords in the actual page content. This means that you need to figure out your keywords before you write any content. Then, you include them throughout your content, particularly in headings and summaries.
So you have heard about microformats, read the introductory articles, and even bought the book. But now you are probably thinking "great - I have done my part to make the web a better place by adding microformats; what's next? What can people do with my data besides add it to their address book or calendar?" The intent of this article is to get you to think about microformats in different ways, and to demonstrate some interesting visualizations and mash-ups of microformatted content.
The Semantic Web is a conceptual information space in which the resources identified by URIs can be processed by machines. It operates on the principles of 'partial understanding' and 'inference' (being able to infer new knowledge of terms from data that you already understand), and hence evolution and transformation. Because the URIs are being used to represent the resources, systems can grow on a globally decentralized basis, similar to hypertext documentation systems on the early WWW.