Computer Science, Option C-Web Science
Computer Science, Option C-Web Science
Computer Science, Option C-Web Science
C.1.2 Describe how the web is constantly evolving
Languages are updated, new specifications and extensions, browsers evolve, increased
bandwidth allows for more, more dynamic, more functions, new programs.
At first web 1.0 just gave information. However, with social media websites users are the ones
that post information. Blogging, tagging, social media has developed the web. Web 3.0 is a
continuation of 2.0 where recommender system like in amazon recommend users items that
other people with the same interest bought.
Describe and compare web 1.0, 2.0 3.0
Web 1.0 is static web pages.
Web 2.0 is dynamic web pages that are driven by user created content, such as Facebook and
Youtube.
Web 3.0 is not clearly defined, but is the idea that the web will become more omnipresent
(internet enabled phones, fridges, cars, etc..) and intelligent (your email will alert you to a
conflict in your calendar with an event that you describe in your email and suggest alternative
dates). There are 3 ways this might happen:
1) Expanded application programming interfaces (APIs) from websites such as
Facebook, that will allow this kind of functionality.
3) The semantic web - web pages are encoded with data (invisible to the user) to add
this functionality. This can be done with a (formal) ontology system or an informal
folksonomy (tagging by users). Currently available ontology systems are ‘rdfa’ and
‘microformat’.
C.1.3 Identify the characteristics of the following:
You can add/remove elements and attributes to the output file. You can also rearrange and sort
elements, perform tests, and decide which elements to hide and display.
JavaScript
Javascript is a scripting language that enables users to make interactive sites. It shares many
features with Java, it has some difference however. Using javascript you can interact with the
HTML source code which can help make dynamic contents.
Javascript is an object based language. Its an interpreting language and not a compiler.
Javascript gives more control to the user over the browser. It can detect the user’s browser and
OS. It performs most of its computations client side rather than server side. Using javascript you
can also validate data, and use it to generate HTML.
https://www.youtube.com/watch?v=Q0k5ySZGPBc
XML is a meta language, that gives meaning to data that other applications can use. Information
about other information in short. Helps information systems share structured data. A meta
language that gives meaning to data that other applications use, application and platform
independent, allows various types of data, extensible to accommodate new tags and processing
methods, allows users to define their own tags.
Simpler version of standardized general markup languages, easy to read and understand,
supported by a large number of platforms, used across open standards.
Youtube: XSLT Example
C.1.4 Identify the characteristics of the following:
● Uniform resource identifier (URI)
● (URL)
URI is the term used for all the types of names/addresses that refer to objects on the world wide
web. A URL is one kind of URI. The URI identifies the object, but does not give a protocol to find
it or a location. For example, a URI could be an ISBN number for a book. It identifies a ‘thing’,
but does not give any information about where to find it.
URL Is the global address of documents and pages/resources on the world wide web. The first
part of a protocol is called a protocol identifier and indicates to the browser what protocol to use.
The second part is called resource name and it specifies the ip address/domain name of where
the resource is located on the world wide web.
C.1.5 Describe the purpose of a URL.
Is the global address of documents and pages/resources on the world wide web. The first part
of a protocol is called a protocol identifier and indicates to the browser what protocol to use. The
second part is called resource name and it specifies the ip address/domain name of where the
resource is located on the world wide web. http:// <- protocol identifier www.google.com = the
location is the resource name
The purpose of URL is to tell the server which webpage to display or to search for. The URL
contains the name protocol to be used to access a file resource. Example of URL:
http://www.google.com
A URL is a formatted text string used by Web browsers, email clients and other software to
identify a network resource on the Internet. Network resources are files that can be plain Web
pages, other text documents, graphics, or programs.
C.1.6 Describe how a domain name server functions.
DNS is a standard technology for managing web sites. DNS allows you to type names into your
Web browser like google.com and turns it into an Internet Protocol (IP) address like
70.42.251.42 that computers use to identify each other on the network.
Understanding DNS:
C.1.7 Identify the characteristics of:
● Internet protocol (IP)
● Transmission control protocol (TCP)
● File transfer protocol (FTP)
Internet Protocol (IP) Is responsible for letting your machine, routers, switches and etcetera,
know where a specific packet is going. IP just mails whatever it gets and can be unreliable as it
does no error checking.
Transmission control protocol (TCP) breaks down data into packets known as TCP segments.
TCP makes data transmission very reliable. TCP recovers from data that is damaged, lost,
duplicated, or delivered out of order by the Network Layer. It achieves multitasking by using port
numbers.
Assigns sequence numbers to to each byte that is transmitted and rearranges TCP segments
into the correct order as they arrive. “Full duplex: TCP provides for concurrent data streams in
both directions”
File Transfer Protocol (FTP) FTP Clients transfer initiates a connection with a remote computer
running FTP server software. AFter that the client can chose to either send and or receive files.
Clients identify FTP servers by its IP address or by its hostname. IT supports two types of data
transfer such as plain text (ASCII) and binary.
C.1.8 Outline the different components of a web page.
Metadata: tags on a page to suggest keywords for search engines.
Title: In the title bar.
Other: Footers, headers, content, navigation bar.
C.1.9 Explain the importance of protocols and standards on the web.
Protocols are a set of rules and procedures that the sender and receiver must adhere to to
ensure coherent data transfer. Examples are TCP, IP, HTTP and FTP.
A standard is an anything that has been agreed upon, which could be a protocol or could be
something like how a network cable is constructed. Most protocols, including the ones listed
above, are also standards, or they would be useless.
Without protocols that are standards, communication networks would not be possible.
C.1.10 Describe the different types of web page.
Personal Pages: A personal webpage is created by an individual for his/her own
personal need.
Blogs: Blogs are basically community forums where you can leave input and is a
discussion or informational site on the world wide web
Search Engine Pages: A search engine results page (SERP) is the listing of results
returned by a search engine in response to a keyword query. The results normally
include a list of items with titles, a reference to the full version, and a short description
showing where the keywords have matched content within the page.
Forums: a place, meeting, or medium where ideas and views on a particular issue can
be exchanged.
C.1.11 Explain the differences between a static web page and a dynamic
web page.
Static web pages are the same every time you visit them.
Dynamic web pages use a database to change depending upon when you visit them (i.e.
Facebook) or what you enter into them. (also Facebook).
C.1.12 Explain the functions of a browser.
The primary functions of web browsers is to provide the resources or information to the user
when asked by them. It processes the user inputs in the form of a URL. it allows the user to
interact with the web pages and dynamic content like surveys and forms. It also provide security
to the data and resources that are available on the web. It also changes HTML into a
displayable format for the user.
Web browser functions are to provide the resources or information the user required and to
display web documents, render HTML and other web formats. It processes the user inputs in
the form of URL like http://www.google.com in the browser and allows the access to that page.
URL is used to identify the resources and fetch them from the server and displays it to the client.
It allows the user to interact with the web pages and dynamic content like surveys, forms, etc. It
also allows the user to navigate through the complete web page and see its source code in the
HTML format. It provides security to the data and the resources by using the secure methods.
C.1.13 Evaluate the use of client-side scripting and server-side scripting in
web pages.
Client side scripts run in the browser and do not require a server to run them. They are
appropriate for processing that does not require any remote information. For example, a good
use would be using javascript to validate a form before sending the information to a server.
Server side scripting are scripts that run on a server. The user never sees the scripts. Think
PHP. Server side scripts are appropriate for processing that is best done on the server, for
example processing a SQL request for a database that is on the server.
C.1.14 Describe how web pages can be connected to underlying data
sources.
Use server side scripts in languages such as PHP to connect connect to and query databases
on a server. The resulting data can then be sent to the web browser.
C.1.15 Describe the function of the common gateway interface (CGI).
The common gateway interface (CGI) is a standard way for a web server to pass a web user's
request to an application program and to receive data back to forward to the user. When the
user requests a web page, the server sends back the requested page.This method or
convention for passing data back and forth between the server and the application is called the
common gateway interface (CGI). It is part of the Web's Hypertext Transfer Protocol.
Websites using CGI usually have a CGI folder with small scripts (typically PERL) that are run on
request. These scripts usually access a database, such as we have been doing with PHP, and
then return the information to the browser that requested it.
C.1.16 Evaluate the structure of different types of web pages.
C.2 Searching The Web
C.2.1 Define the term search engine
Search engines are programs that search documents for specified keywords and returns a list of
the documents where the keywords were found.
C.2.2 Distinguish between the surface web and the deep web.
https://www.youtube.com/watch?v=_UOK7aRmUtw
The deep web is any website that cannot be indexed by a search engine. Only 10% of the web
is indexed by search engines. The deep web is basically anything you have to log into. It is the
part of the web where access is restricted unless a member of that website for example the
deep web can consist of your bank account portal where the general public doesn’t have access
to that information. The surface web is what we see when we use a search engine. Basically
search engines send out bots regarding your search topic and then displays them. Those links
generally appear on the surface or can easily be found.
C.2.3 Outline the principles of searching algorithms used by search engines.
A good comparison of HITS and pagerank, just watch it up until 2:13 ---
C.2.4 Describe how a web crawler functions.
https://www.youtube.com/watch?v=CDXOcvUNBaA A web crawler is an internet bot that
systematically browses the the world wide web, typically for the purpose of web indexing. Web
indexing refers to various methods for indexing the contents of a website or of the internet as a
whole. As web crawlers visits URLs, it identifies all the hyperlinks in the page and adds them to
the list of URLs to visit in a list and then displays the results.
C.2.5 Discuss the relationship between data in a meta-tag and how it is
accessed by a web crawler.
The meta tag provides metadata about the HTML document. Metadata is data that describes
and gives information about other data. Metadata will not be displayed on the page, but will be
machine parsable. Web crawlers when sent out to find a specific webpage look at the meta data
and determine whether the description of the web page is sufficient enough to be shown as a
search result. If not, the crawler doesn’t compile that specific link in the list of results pertaining
to the search.
IB Spec
#2c The organizers of the theatre want to ensure that their web pages appear higher up
the ranking of search engines. Suggest whether the use of meta-tags can help achieve
this aim. (4)
C.2.6 Discuss the use of parallel web crawling
A parallel crawler is a crawler that runs multiple processes in parallel. The goal is to maximize
the download rate of information. To avoid downloading the same page more than once, the
crawling system requires a policy which states that if the same link is found by two separate
crawlers, the second one should disregard it.
C.2.7 Outline the purpose of web-indexing crawling in search engines.
A search engine would likely store more than just the keywords and url that the user search for;
they would also store the number of times the words appear on that page.Web-indexing is when
the crawlers called spiders, take the keywords in the search box input there by the users and
search the web for the pages that contains the keywords, synonyms or links that will link to what
users are looking for.
Youtube video:
Link: https://www.google.com/insidesearch/howsearchworks/crawling-indexing.html
C.2.8 Suggest how web developers can create pages that appear more
prominently in search engine results.
The ideal optimized web page has a clear editorial content focus, with the keywords or phrases
present in these elements of the page, in order of importance:
● Page titles: titles are the most important element in page topic relevance; they also have
another important role in the web interface—the page title becomes the text of
bookmarks users make for your page
● Major headings at the top of the page (<h1>, <h2>)
● The text of links to other pages: link text tells a search engine that the linked words are
important; “Click here” gives no information about the destination content, is a poor
practice for markup and universal usability, and contributes nothing to the topical
information of the page
● The alternate text for any relevant images on the page: accurate, carefully crafted
alternate text for images is essential to universal usability and is also important if you
want your images to appear in image search results such as Google Images and Yahoo!
Images; the alternate text of the image link is the primary way search engines judge the
topical relevance of an image
● The HTML file name and the names of directories in your site: for example, the ideal
name for a page on Bengal tigers is “bengal-tigers.html”
C.2.9 Describe the different metrics used by search engines.
Pagerank score
PageRank works by counting the number and quality of links to a page to determine a rough
estimate of how important the website is. More important websites are likely to receive more
links from other websites
Hub score
A page is given a high hub score by linking to nodes that are considered authorities on the
subject
C.2.10 Explain why the effectiveness of a search engine is determined by
the assumptions made when developing it.
Search engines are programmed to search based on a fixed algorithm and ranking order
making general assumptions about what the user is looking for, regardless of the user’s intent or
circumstances
For example Google’s pagerank ranking order makes the assumption that if more sites link to
you, you are a more valuable site. HITS ranking order assumes that if you are linking to other
sites you are a valuable site yourself, regardless of content.
The same search may be valuable to one user while useless to the next
To improve this quality of searches some superficial features have been added, such as custom
google searches, or user preferences in settings. However these do not solve the problem of
the search engine lacking insight into the semantics of what the user wants
A way the search engine could perform a better search is with follow up questions, such as if
you search for a camera, it asks whether you are interested in comparing models and prices,
interested in purchasing one, or interested about the general art of photography
This information could be used later to allow the engine to make better assumptions about the
user and provide a better search
SOURCE
C.2.11 Discuss the use of white hat and black hat search engine
optimization.
Manipulates the search system in order to Caters to human search techniques to make
make a page show up more often in searches it easy to find the page when looked for
Sites are often filled with links and ads Sites may feature relevant multimedia
(videos, etc)
Sites generally don’t look very appealing (the Sites are often very visually pleasing (the
purpose is to get users to just see the purpose is to attract users who are genuinely
content) interested in the page content)
Image:
http://www.prolificwebsolutions.com/wp-content/uploads/black-hat-white-hat-seo-comparison.jpg
C.2.12 Outline future challenges to search engines as the web continues to
grow.
Challenges to future search engines:
● Error management
● Lack of quality assurance of info uploaded
C.3 Distributed Approaches To The Web
C.3.1 Define the terms: mobile computing, ubiquitous computing,
peer-2-peer network, grid computing.
Mobile Computing - Mobile computing is human–computer interaction by which a computer is
expected to be transported during normal usage. Mobile computing involves mobile
communication, mobile hardware, and mobile software. (i.e. laptops and smartphones, smart
cards (normally for payment or transport) and wearable computers like smart watches)
Peer-2-Peer Network - A network of computers configured to allow certain files and folders to
be shared with everyone or with selected users. Peer-to-peer networks are quite common in
small offices that do not use a dedicated file server. Equal clients without a server.
Grid Computing - Grid computing is the collection of computer resources from multiple
locations to reach a common goal. The grid can be thought of as a distributed system with
non-interactive workloads that involve a large number of files. Has a server that is coordinating.
Usually used to solve technical problems, such as SETI.
C.3.2 Compare the major features of:
● Mobile Computing - Mobile computing is human–computer interaction by which a
computer is expected to be transported during normal usage. Mobile computing involves
mobile communication, mobile hardware, and mobile software.
● Ubiquitous Computing- Ubiquitous computing is a concept in software engineering and
computer science where computing is made to appear everywhere and anywhere. In
contrast to desktop computing, ubiquitous computing can occur using any device, in any
location, and in any format.
● peer-2-peer network- P2P is a category of distributed system which optimizes the
benefits of resources such as storage or processing time that are available on the
internet. Implementing P2P requires the creation of overlay networks
● Grid Computing is a type of distributed computing which permits and ensures the
sharing of aggregated resources across dispersed locations. Resources are connected
through a required middleware such as Globus, Legion, or gLite with the internet to
provide various services for management of resources and security issue. Grid
computing is limited to tasks that have no time dependency, since all nodes are working
independently and many are likely to connect and disconnect from the network in any
given time period. For tasks that are highly time independent - since failure on one node
cannot effect another node.
The point of peer to peer networks is to share resources (i.e. bittorrent). The point of grid
computing is to solve problems (i.e. SETI).
C.3.3 Distinguish between interoperability and open standards.
Interoperability - The ability of software and hardware on different machines from different
people or companies to share data.
C.3.4 Describe the range of hardware used by distributed networks.
Grid Computing: Desktop computers running middleware to coordinate them. A network
connection such as the internet.
Ubiquitous computing: Small processors and input/output devices, wireless data connections
such as wifi or bluetooth. All kinds of sensors.
Parallel computing: Fast and small networks such as fiber optics and powerful computers with
specialized operating systems to coordinate highly time dependent tasks.
*********************************************
C.3.5 Explain why distributed systems may act as a catalyst to a greater
decentralization of the web.
Distributed system is an application that executes a collection of protocols to coordinate the
actions of multiple processes on a network. In other words, it is to perform a single or small set
of related tasks. (Google Code)
Decentralization of the web allows for remote users to connect to the remote resources in an
open and scalable way. Distributed system can be a larger and more powerful given the
combined capabilities of the distributed system. This also provides easier access for multiple
people at once instead of only working on the centralized one.
C.3.6 Distinguish between lossless and lossy compression.
In lossless compression, every single bit of data that was originally in the file is kept after the file
is uncompressed. Essentially, all the information is restored.
Lossy compression is when a file is permanently reduced (eliminate some data). It usually
eliminated redundant information. When it is uncompressed only part of the original file will be
there. Commonly used in pictures, sound, video where some loss in quality would be hard to
identify.
For example, a picture stored in Raw format records the color for every pixel in an image. This
is (mostly) lossless. The files are very large. A jpg compresses this information and makes the
image much smaller, but loses some quality.
Answer:
C.3.7 Evaluate the use of decompression software in the transfer of
information.
Compression software is essential to lowering file size. If file size is smaller, it gives users that
attempt to open website faster loading times. Which essentially makes the website more
responsive. Mobile users that open files on their phones will experience lower costs when files
are compressed.
C.4 The Evolving Web
To answer this, we must first look at the stakeholders of the Web:
● S/holder of the web: Developers, Administrators, Governments
● S/holder of companies: Customers,Vendors, Employees
Online Interaction has been developed through the different iterations of the Web :
● Web 1.0
● Web 2.0
● Web 3.0 ( If it exists )
Web 1.0 - Static websites - Close to NO interaction, as users do not have an input
Web 2.0 - User to website model - users have the ability to submit information to a website, so
the level of interaction is greater - however, there is not a full integration of user input onto the
web, as the interaction is often done through a website. Examples include Facebook and
Instagram, which allows users to communicate and share information through these websites.
Web 3.0 - Although the existence of Web 3.0 is questioned, the idea is that interaction is user
to user, with no interference from websites. An example would be if there are many low ratings
for a restaurant on a review site such as hellofood.com, the restaurant/website may delete the
reviews, whereas in web 3.0, users would be able to input “real” ratings without interference.
TO Summarize:
– Web 1.0: No interaction between computer and users.
– Web 2.0: Interaction between both computers and users.
– Web 3.0: Real ratings, true interactions between users without websites interfering.
C.4.3 Discuss the effects of the use of cloud computing for specified
organizations.
Cloud computing is appealing to organizations due to its lower costs, less need for space, and
natural destruction recovery measures. Cloud computing allows users in the organization to
connect and share files and data with one another.
Public:
Public cloud computing is low cost and is based upon infrastructure and servers located
externally that are already put in place. This makes it easy to transition to public cloud
computing and saves time and money. However due to the fact that the servers and information
are being stored externally, and other users may be using the same cloud computing services,
security is an issue as there is no way to secure the information other than the protocols already
put in place by the cloud provider.
Private:
Private cloud computing has been gaining traction amongst organizations. Private cloud
computing is built on a locally hosted server and computer network, so it requires more cost to
set up. However these organizations will likely already have a server and local network setup so
a private cloud computing network would be a return on the company's investment. The storage
and effectiveness of the computing system is reliant on the hardware that the company owns,
yet a private cloud system is much more secure as the information and data is stored locally.
The company still maintains the advantages of cloud computing, such as the destruction
recovery.
IB Questions:
Specimen Paper C1:
May 2015 11D
C.4.4 Discuss the management of issues such as copyright and intellectual
property on the web.
Issue concerning intellectual property on the web occur when the owners rights collide with
users rights and the public need to access and use resources on the web.
Copyright is a type of intellectual property, giving your rights to copy and distribute the relevant
intellectual property. However, you might have intellectual property that you do not have
copyright (legal rights) to. i.e. an idea for an invention that you have not yet copyrighted.
Authentication is the verification of the claimed identity. An example would be checking the
database to find a match between the username and password entered by the user, and if a
match is found they are granted entry to the site.
Privacy is the safekeeping of identification. For example, most websites that require a login are
private, because your password and/or username are not displayed for others to see. However,
some wifi networks are not private, and other computers on the network may be able to see
your IP address.
(d) Comment on the privacy and security issues relating to ABC’s use of cloud computing.
Markscheme:
Award marks as follows up to [4 marks max].
Award [2 marks] for a privacy issue identified and elaborated;
Award [2 marks] for a security issue identified and elaborated;
Privacy
Sensitive data is accessible to a third party;
If outsourcing occurs, potential exposure of data is increased;
Security
How secure is the data?;
Can it be guaranteed that this data will not be inadvertently passed to another company?;
Short Youtube video (< 4mins) detailing relationship between identification and authentication:
https://www.youtube.com/watch?v=bv6QL0H8kAY
C.4.6 Describe the role of network architecture, protocols and standards in
the future development of the web.
All 3 must develop to support increasing traffic and new developments such as ubiquitous
computing, cloud services and other new developments that we cannot anticipate now. For
example,a new IP system is needed to support more IP addresses (for example, IP 6). New
network architecture will be needed to support ever more devices and speeds. This new
architecture will require new protocols and standards.
C.4.7 Explain why the web may be creating unregulated monopolies.
Often, as a website grows larger, it becomes more dominant and more difficult for rivals to
challenge. For example, everyone joins Facebook because that’s what most people use. It’s
hard to start a rival because networks with less users are less useful. It’s hard to to challenge
Amazon’s dominance in e-commerce because sellers use it because it has the most customers
and customers use it because it has the most sellers. It’s scale allows it to negotiate very
favorable terms with shipping companies and suppliers.
C.4.8 Discuss the effects of a decentralized and democratic web.
Decentralized refers to the idea that content creation can come from anywhere, which promotes
participation (democracy). Note that in this context, democratic has nothing to do with politics, it
just means everyone can participate.
C.5 Analysing the Web
The graphic below is linked to a longish but good video about all the topics in this section.
C.5.2 Outline the difference between the web graph and sub-graphs.
A subgraph is a portion of a web graph that only shows part of the web. Whilst, a web graph is
the connection of all the pages on the internet a sub graph only shows a portion of the
connections. For example, the connections at ISK.
C.5.3 Describe the main features of
the web graph such as bowtie
structure, strongly connected core
(SCC), diameter.
A bow-tie map of the internet offers a three
dimensional model of understanding the web.
Reveals a subtler structure that may lead to
more efficient search engine crawling
techniques. The bow tie structure resembles a
bow tie. Which consists of three major regions a.
knot and two bows, and a fourth structure of the
pages that are disconnected.
The left bow consists of pages (origination) that allow users to reach the core. However, at the
same time they can’t be reached from the core. Origination pages, are pages that haven’t
attracted enough attention from the internet. Such that no pages in the core link to them. These
sites usually have links that link to each other but don’t reach the core of the internet.
The right bow consists of pages that can be accessed from the core but don’t link back to the
core. Pages in the right bow only contain information about themselves.
Disconnected pages are not part of the bow tie. These pages can connect to either the
origination/termination pages but can’t access the core directly.
SCC and diameter
C.5.4 Explain the role of graph theory in determining the connectivity of
the web.
Graph theory is essential in learning and determining how the web is connected. For example,
using bow tie theory. You can learn how web pages are interconnected and how sites are
interconnected to reach the core.
C.5.5 Explain that search engines and web crawling use the web graph to
access information.
Search engines and web crawlers use web graphs to determine the importance of information.
Currently the web only display 40-70% of all the information on the internet. As such not all of
the information is being shown. Web crawlers only identify links that have relevant data and
quality. The websites popularity is also determined depending on the amount of links and visits.
Using bow tie theory to determine how we can see how disconnected pages and origination
pages will not be accessed by web crawlers.
C.5.6 Discuss whether power laws are appropriate to predict the
development of the web.
(The rate at which the internet grows) Power laws suggest that the internet is developing
exponentially. However, it seems unlikely that exponential growth can continue indefinitely.
The network diameter of the web growing no more than logarithmically with respect to
the network size. Sources suggest there were 26 million pages in 1998, one billion
pages by 2000 and 34 billion by 2011;
Therefore the diameter is not growing in a linear relationship with the number of web
pages, so the ability of the web surfer to access all sites remains possible;
10-fold increase of web pages results in only 2 more additional “clicks”;
C.6 The Intelligent Web
C.6.1 Define the term semantic web.
Semantics is related to syntax. Syntax is how you say something and semantics is the meaning
behind what you said. Although syntax can change, the meaning behind the syntax can stay the
same. For example, “I love you” and “I heart you” have different syntax (grammar), but the
same semantics (meaning).
Computers don’t understand the content of web pages, however they can understand the
syntax. Help computers understand the meaning of the website. Web 2.0 is documents, and
Semantic is things. Pointing things out to the computer and letting them know what they mean.
If a computer can understand what is going on it can help you in many ways.
The ‘semantic web’ is the idea of using emerging technologies such as ‘microformats’ to store
information about all kinds of objects on the web, and how they interconnect. So, reservations,
can be connected to calendars, which can be connected to recommendations, etc…
C.6.2 Distinguish between the text-web and the multimedia-web.
Text-web is a web browser that renders only the text of web pages, and ignores graphic
content. Usually, they render pages faster than graphical web browsers due to lowered
bandwidth demands. Multimedia refers to content that uses a combination of different content
forms. Unlike text-web, multimedia web pages use different forms of graphic content.
C.6.3 Describe the aims of the semantic web.
The aim of the semantic web is to evolve the current web by enabling users find, share, and
combine information more easily. What the semantic web is aiming to be is a vision of
information that can be interpreted by machines so they can perform more tedious tasks
involved with finding, sharing and combining information.
The purpose of the semantic web is to enhance the usability and usefulness of the web and its
resources by creating semantic web services such as:
https://www.youtube.com/watch?v=bEYQrmPwjPA
C.6.4 Distinguish between an ontology and folksonomy.
A folksonomy is a distributed form of tagging - users decide on tags themselves.
An ontology is a standardized form of tagging - a group or organization agrees on the tags that
are possible - this makes it easier to merge and process data, but is more difficult to achieve.
Ontology is a formal naming and definition of the types, properties, and interrelationships of the
entities that really or fundamentally exist for a particular domain of discourse. Folksonomy is a
system of classification derived from the practice and method of collaboratively creating and
translating tags to annotate and categorize content.Ontology is hard to implement on a large
scale and isn’t always web based. It is Key for the semantic web because of high expressive
power. Folksonomy is created by users as well as being quick and easy to implement. It is used
on a large scale for document collections. Most of the time its web based and important in web
2.0
http://www.cl.cam.ac.uk/~aac10/R207/ontology_vs_folksonomy.pdf
An ontology is the formal description of the concepts and relationships that exist
within a specified domain of discourse.
“A folksonomy is a type of distributed classification system. It is usually created
by a group of individuals, typically the resource users. Users add tags to online
items, such as images, videos, bookmarks and text. These tags are then shared
and sometimes refined.”
An example of folksonomies is the tag system in image sites or hashtags in social media. Users
are defining more and more tags and as the volume of users tagging increases, the accuracy of
tags increases such that the web is becoming more and more precise.
Tagging increases user participation in the web while enhancing searching and semantics of the
web.
C.6.6 Explain why there needs to be a balance between expressivity and
usability on the semantic web.
With the idea of using “ Meta-tagging” as a part of the semantic web, there comes the problem
of find a common ground between usability and expressivity - even with the idea that a
computer can understand the information that is on a page, there needs to be a balance to
achieve a level where the expressivity does not compromise the usefulness that is intrinsic to
proper syntax or language.
Without a balance, the idea of semantics becomes inefficient - if there is a lack of structure to
the way data concerning the semantics of a page is formatted, It makes it difficult for search
engines and other applications to understand and make use of this information - Hence, it is
necessary to define a language or structure by which the semantics of a webpage or application
C.6.7 Evaluate methods of searching for information on the web.
Some methods of information searching on the web are using unique and specific terms, using
filters to narrow the search, and custom searches. By using specific terms that pertain to your
search, you reduce the number of results that display. By using filters, you’re able to find your
information based on relevance, date added so on and so forth. Custom searches also limit the
number of results by using different operators.
C.6.8 Distinguish between ambient intelligence and collective intelligence.
Ambient intelligence refers to electronic environments that are responsive to the presence of
people. Ambient intelligence is aimed at supporting people in their daily lives and activities.
Technology that becomes small enough to be considered a part of the environment.
Youtube about ambient intelligence: https://www.youtube.com/watch?v=dy-QKM3i36E
Collective intelligence refers to intelligence shared by collaboration such that many people are
involved in consensus decision making. For example a google doc with many collaborators
inputting ideas, wikipedia, or using a website to help people collectively analyze and discuss
climate change.
Youtube about collective intelligence: https://www.youtube.com/watch?v=k7-CEDyoibQ
IB 2014 P2
#4f: Identify two characteristics of collective intelligence. (2 marks)
Answer:
C.6.9 Discuss how ambient intelligence can be used to support people.
Ambient Intelligence (AmI) is bringing technology into our environment and enabling for the
environment to become sensitive to us. Researchers build upon advances in sensors and
sensor networks, pervasive computing and artificial intelligence in order to make the
environment more aware of us and react accordingly if necessary.
Biometrics is an example that can be used to identify people. In India, there are many people
without identification and can’t use banking services or microcredit loans. Having everyone
tagged with biometrics will allow them to use these services. Another way ambient intelligence
can support people is if culprit x is identified by an intelligent surveillance camera, it will
immediately contact the police and inform them of his location.
Reading: How AmI can help with Health Services
Reading: Further information about Ambient Technology
Video: The Future of Ambient Life
Complex issues, such as climate change, can be addressed using methods of collective
intelligence. One example is the monitoring of temperature across the globe. Weather probes
located in airports, television stations, and more, automatically transmit data gathered about the
temperature to a central source such as the National Weather Service (NWS) in the US. The
NWS then uses this data to make predictions, graph weather patterns, and more. This is an
example of technology being used in collective intelligence to solve complex issues, which
would be made much more difficult without the benefit of technology.
Picture: http://www.henshall.com/blog/archives/images/mainresources-thumb.jpeg