The future of the Web is a matter of semantics
The first incarnation of the web was composed of static websites that linked to each other and search engines to help you find sites of interest. Web 2.0 brought a social element to the web, with users sharing, commenting, and interacting through sites such as YouTube, Facebook, and Flickr. The future web, the “semantic web”, or Web 3.0, will embed meaning within digital information so that any given page can be understood by computers as well as people.
The inventor of the World Wide Web, Tim Berners-Lee first mentioned the concept of a semantic web, a web with inbuilt meaning, long before the advent of social sites, but it is yet to become reality. This is despite the ongoing efforts of web engineers, academics, search engine companies, and the web industry itself. There is, researchers writing in the International Journal of Web Engineering and Technology, a semantic web bottleneck.
Nikolaos Konstantinou of Athens Information Technology (AIT) and colleagues at the National Technical University of Athens (NTUA), in Greece, state that after almost a decade of research, the fundamental concepts that would underpin a semantic web have matured, yet the average web user cannot yet take advantage of their full potential. They suggest that there are three main issues to be overcome before Web 3.0 emerges and they present a roadmap in their paper to explain how these must be addressed.
In Berners-Lee’s original vision for the semantic web, machine-readable information embedded in a digital object, whether a web page, an image, a video or some other file, so-called meta data, would allow software to potentially understand the meaning and context of the digital object. Although some software currently has a limited understanding of simple meta data, it mostly lies in prototypes and lab environments.
However, the potential of the semantic web is to have software agents that can perform tasks automatically based perhaps on a user’s behaviour or preference settings, and to locate pertinent information far more efficiently than an individual searching the web manually might do. The software might also be able to infer additional knowledge based on previously existing information process the information it finds into a usefully organised format. Such a process would be useful to scholars, doctors, engineers, scientists, musicians, designers, artists, indeed anyone who works with data.
Konstantinou and colleagues point out that three issues are preventing this from happening: a lack of simplicity, integration with existing technologies and practices, and adoption by the web industry.
They suggest that ways to automatically add meta data to digital objects are now needed to make it possible to publish semantically rich content without manual intervention regardless of whether the “publisher” is a large corporation or an individual content creator. They also say that semantic technologies do not offer a substitute for current practices, rather a complement to them and that web engineers need not abandon experience but should build on it. Finally, the driving forces of the web industry should adopt semantic web technologies since their adoption entails a series of benefits both for the companies themselves as well as to the end users. “This seems to be the most promising solution for the chicken-and-egg problem of the semantic web,” the team says. “Much still needs to be done in order to effectively publish and exploit large-scale semantic information. Following the approach suggested in this paper, we are confident that the semantic web bottleneck will be shortly circumvented and the semantic web vision will be at last realised,” the team concludes.