Books, Brochures, and Chapters>Book:  Berners-Lee, Tim (2006-09-15), A Framework for Web Science (Foundations and Trends(R) in Web Science), Now Publishers Inc, Retrieved on 2010-11-15
  • Source Material [eprints.ecs.soton.ac.uk]
  • Folksonomies: web science

    Memes

    02 JAN 2011

     Page Rank Algorithm and the Kleinberg's HITS algorithm

    The most famous quality measure is PageRank [221], discussed earlier, which builds on the intuition that a page which is cited by many other pages is likely to be of significant quality. The insight of PageRank is that the obvious way to subvert that model is to set up a load of dummy pages to cite the page which one wanted to boost. But if a page is cited by many other pages which themselves have a high PageRank, then it is likely to be of high quality. The PageRank method has another intuit...
      1  notes

    The page rank algorithm is very similar to an algorithm for ranking research papers by counting the number of other papers referencing them.

    02 JAN 2011

     Social Goals of Web Science

    There are two important epistemological questions for Web Science. The first is what properties will future platforms need to have in order to allow as much information as possible to gravitate to the Web without imposing structure or governing theories upon it? One aim of the Web is to facilitate rational discussion of ideas, rather than the sorts of rancorous ad hominem attacks that make up rather too much of what is loosely called debate [30]. And secondly, the Web has a radically decen...
      1  notes

    Epistemological questions for Web Science. Such as providing an online environment that allows rational discussion without imposing structure on it.

    02 JAN 2011

     Associative vs. Analogous Reasoning

    There are many different types of reasoning, but not too many have been successfully automated beyond deductive linear reasoning and various statistical methods. What alternative methods has the Web facilitated? One obvious candidate is associative reasoning, where reasoning on the basis of associations - which can be extremely unpredictable and personalized - takes one down a train of thought [202]. So, for example, the classic case of associative reasoning is given in Proust's novel Remembr...
      1  notes

    How information technologies aid these two different forms of reasoning.

    02 JAN 2011

     The Concept of Supervenience and the Web

    One view is reminiscent of the philosophical idea of supervenience [168, 169]). One discourse or set of expressions A supervenes on another set B when a change in A entails a change in B but not vice versa. So, on a supervenience theory of the mind/brain, any change in mental state entails some change in brain state, but a change in brain state need not necessarily result in a change in mental state. Supervenience is a less strong concept than reduction (a reductionist theory of the mind/brai...
      1  notes

    Where changes in one concept cascade into changes on another, but not vice versa.

    02 JAN 2011

     Information Retrieval is an Arms Race Between Algorithms ...

    IR is the focus for an arms race between algorithms to extract information from repositories as those repositories get larger and more complex, and users' demands get harder to satisfy (either in terms of response time or complexity of query). One obvious issue with respect to IR over the Web is that the Web has no QA authority. Anyone with an ISP account can place a page on the Web, and as is well known the Web has been the site of a proliferation of conspiracy theories, urban legends, tr...
      1  notes

    As the web gets larger and data grows more complex, less trustworthy in many regards, algorithms will need to grow more sophisticated to adapt to it.

    02 JAN 2011

     Compactness in Linked Hypertext

    Global metrics look at extracting information about the graph as a whole. Compactness is a measure of how connected the graph is; a compact graph means that, in general, it is easy to reach a randomlychosen node from another. The usual measure has a range between 0 (totally disconnected nodes) and 1 (universal connections). Compactness of 0 is obviously hopeless for an information space, but perhaps less obviously the graph shouldn't be too compact either; if authors of webpages are sparing a...
      1  notes

    A brief summary of compactness in understanding web topology and balance.

    02 JAN 2011

     Graph Theory Approach to Web Topology

    Perhaps the best-known paradigm for studying the Web is graph theory. The Web can be seen as a graph whose nodes are pages and whose (directed) edges are links. Because very few weblinks are random, it is clear that the edges of the graph encode much structure that is seen by designers and authors of content as important. Strongly connected parts of the webgraph correspond to what are called cybercommunities and early investigations, for example by Kumar et al, led to the discovery and mappin...
      1  notes

    The graph theory approach produces a model of the web that is like a bowtie, and filled with other bowties, like a fractal. There is an image in the original document of this phenomena.

    02 JAN 2011

     The Importance of Web Topology

    Web topology contains more complexity than simple linear chains. In this section, we will discuss attempts to measure the global structure of the Web, and how individual webpages fit into that context. Are there interesting representations that define or suggest important properties? For example, might it be possible to map knowledge on theWeb? Such a map might allow the possibility of understanding online communities, or to engage in 'plume tracing' - following a meme, or idea, or rumour, or...
      1  notes

    Mapping the web allows us to find patterns in it, with potential applications.

    02 JAN 2011

     The Web is More Than Text

    Nevertheless, the next generation Web should not be based on the false assumption that text is predominant and keyword-based search will be adequate for all reasonable purposes [127]. Indeed, the issues relating to navigation through multimedia repositories such as video archives and through theWeb are not unrelated: both need information links to support browsing, and both need engines to support manual link traversal. However, the keyword approach may falter in the multimedia context becaus...
      1  notes

    All of our search technologies, semantic explorations, and other online conversations are all dependent on text, but they must grow to be able to read images, audio, and video as well.

    02 JAN 2011

     Assessing Trust, Subjective Logic, Real-Valued Triple

    Another key factor in assessing the trustworthiness of a document is the reliability or otherwise of the claims expressed within it; metadata about provenance will no doubt help in such judgements but need not necessarily resolve them. Representing confidence in reliability has always been difficult in epistemic logics. In the context of knowledge representation approaches include: subjective logic, which represents an opinion as a real-valued triple (belief, disbelief, uncertainty) where the...
      1  notes

    Need to follow up on these concepts, learn more about them for establishing the trustworthiness and quality of content.

    02 JAN 2011

     The Importance of Time-Stamping to Relevance

    Time-stamping is of interest because the temporal element of context is essential for understanding a text (to take an obvious example, when reading a paper on global geopolitics in 2006 it is essential to know whether it was written before or after 11th September, 2001). Furthermore, some information has a 'sell-by date': after a certain point it may become unreliable. Often this point isn't predictable exactly, but broad indications can be given; naturally much depends on whether the inform...
      1  notes

    Time-stamping is a crucial function of semantic data. Some information grows less accurate over time, while a context is desired for other data.

    02 JAN 2011

     Ontologies vs. Folksonomies

    It is argued - though currently the arguments are filtering only slowly into the academic literature - that folksonomies are preferable to the use of controlled, centralised ontologies [e.g. 259]. Annotating Web pages using controlled vocabularies will improve the chances of one's page turning up on the 'right' Web searches, but on the other hand the large heterogeneous user base of the Web is unlikely to contain many people (or organisations) willing to adopt or maintain a complex ontology. ...
      1  notes

    Ontologies provide structure and a standard for tagging and searching, while folksonomies provide for an emergent system for tagging things.

    02 JAN 2011

     Natural Language Processing vs. Semantic Web

    NLP works well statistically; the SW, in contrast, requires logic and doesn't yet make substantial use of statistics. Natural language is democratic, as expressed in the slogan 'meaning is use' (see Section 5.1 for more discussion of this). The equivalent in the SW of the words of natural language are logical terms, of which URIs are prominent. Thus we have an immediate disanalogy between NLP and the SW, which is that URIs, unlike words, have owners, and so can be regulated. That is not to sa...
      1  notes

    A short comparison of the difference between NLP and SW in terms of processing, algorithms, structure, and emergence. NLP is described as 'democratic', where the power of SW is that URIs 'have owners,' meaning they are a top-down construct. Perhaps this is the problem of the Semantic Web and why it may never catch on: the web favors emergent semantics and democratized development.

    02 JAN 2011

     The Web as a Liberal Artefact

    A final point briefly worth making is that the Web is a space designed to let information flow, and to create opportunities for cooperation and collaboration. It is worth asking why freer information flow is a good thing, and the answers are pretty straightforward. It is good to have the freedom to express oneself in order that one can pursue one's own autonomous and authentic projects. Unhindered criticism of governments and other power centres tends to lead to better governance; information...
      1  notes

    With openess, egalitarianism of ideas, and free flow of information as its principles, the Web works as a liberal democracy, and totalitarian governments attempt to control or subvert it.

    Parent Reference