CERN Accelerating science

home

The birth of the Web

The World Wide Web was invented by British scientist Tim Berners-Lee in 1989 while working at CERN

Tim Berners-Lee, a British scientist, invented the World Wide Web (WWW) in 1989, while working at CERN. The web was originally conceived and developed to meet the demand for automated information-sharing between scientists in universities and institutes around the world.

home.cern,Accelerators

The first website at CERN – and in the world – was dedicated to the World Wide Web project itself and was hosted on Berners-Lee's NeXT computer. In 2013, CERN launched a project to restore this first ever website :  info.cern.ch .

On 30 April 1993, CERN put the World Wide Web software in the public domain. Later, CERN made a release available with an open licence, a more sure way to maximise its dissemination. These actions allowed the web to flourish.

A short history of the Web

Licensing the web, browse the first website, the worldwideweb browser, the line-mode browser, the cern effect, internet prehistory at cern, minimising the muddle, good old bitnet, and the rise of the world wi..., on the open internet and the free web, not at all vague and much more than exciting, why bring back the line-mode browser, how the internet came to cern, twenty years of a free and open www.

assignment on world wide web

Advertisement

What's the Difference Between the Internet and the World Wide Web?

  • Share Content on Facebook
  • Share Content on LinkedIn
  • Share Content on Flipboard
  • Share Content on Reddit
  • Share Content via Email

Imogen Heap, Web summit

The Internet has become so ubiquitous it's hard to imagine life without it. It's equally hard to imagine a world where "www" isn't the prefix of many of our online activities. But just because the Internet and the World Wide Web are firmly intertwined with each other, it doesn't mean they're synonymous.

Let's go back to when it all began. President Dwight D. Eisenhower started the Advanced Research Projects Agency (ARPA) in 1958 to increase U.S. technological advancements in the shadow of Sputnik's launch. By October 29, 1969, the first ARPANET network connection between two computers was launched — and promptly crashed. But happily, the second time around was much more successful and the Internet was born. More and more computers were added to this ever-increasing network and the megalith we know today as the Internet began to form. Further information about ARPA can be discovered by reading How ARPANET Works .

But the creation of the World Wide Web didn't come until decades later, with the help of a man named Tim Berners-Lee. In 1990, he developed the backbone of the World Wide Web — the hypertext transfer protocol ( HTTP ). People quickly developed browsers which supported the use of HTTP and with that the popularity of computers skyrocketed. In the 20 years during which ARPANET ruled the Internet, the worldwide network grew from four computers to more than 300,000. By 1992, more than a million computers were connected — only two years after HTTP was developed [source: Computer History Museum ].

You might be wondering at this point what exactly HTTP is — it's simply the widely used set of rules for how files and other information are transferred between computers. So what Berners-Lee did, in essence, was determine how computers would communicate with one another. For instance, HTTP would've come into play if you clicked the source link in the last paragraph or if you typed the https://www.howstuffworks.com URL ( uniform resource locator ) into your browser to get to our home page. But don't get this confused with Web page programming languages like HTML and XHTML . We use those to describe what's on a page, not to communicate between sites or identify a Web page's location.

To learn more about the dawn of the Internet age, visit How did the Internet start? For our purposes, we're set to explore the fundamental difference between the Internet and the World Wide Web, and why it's so easy for us to link them together in our minds. Go to the next page to find the answers.

Internet vs. World Wide Web

Cannes Lions International Festival, Tim Bernes-Lee

To answer this question, let's look at each element. And since the Internet seems to be the more easily understood component, let's start there.

Simply, the Internet is a network of networks — and there are all kinds of networks in all kinds of sizes. You may have a computer network at your work, at your school or even one at your house. These networks are often connected to each other in different configurations, which is how you get groupings such as local area networks ( LAN s) and regional networks . Your cell phone is also on a network that is considered part of the Internet, as are many of your other electronic devices. And all these separate networks — added together — are what constitute the Internet. Even satellites are connected to the Internet.

The World Wide Web, on the other hand, is the system we use to access the Internet. The Web isn't the only system out there, but it's the most popular and widely used. (Examples of ways to access the Internet without using HTTP include e-mail and instant messaging .) As mentioned on the previous page, the World Wide Web makes use of hypertext to access the various forms of information available on the world's different networks. This allows people all over the world to share knowledge and opinions. We typically access the Web through browsers, like Internet Explorer, Chrome and Mozilla Firefox . By using browsers like these, you can visit various Web sites and view other online content.

So another way to think about it is to say the Internet is composed of the machines, hardware and data; and the World Wide Web is what brings this technology to life.

Now that we know the difference between the Internet and the World Wide Web, put your newfound knowledge of hyperlinks, HTML and home pages to use and click onto the next page for more great information.

There are several groups of people who work to keep everything standardized and running smoothly across the Internet and the World Wide Web. If you want to learn more about what it takes to keep systems compatible, a good place to start is the World Wide Web Consortium ( W3C ). W3C's aim is to help develop standards and guidelines for the Web. The group is run by Tim Berners-Lee who, if you remember, is the person who invented the World Wide Web. You can also try the Internet Society , which was founded in 1992 to provide leadership in many aspects of Internet-related information, initiatives and activities. 

Lots More Information

Related howstuffworks articles.

  • How ARPANET Works
  • How Bits and Bytes Work
  • How did the Internet start?
  • How Internet Infrastructure Works
  • How Semantic Web Works
  • How Web Pages Work

More Great Links

  • Computer History Museum
  • Defense Advanced Research Projects Agency
  • Internet Corporation for Assigned Names and Numbers
  • Internet Network Information Center
  • National Science Foundation
  • Berners-Lee, Tim. "Frequently Asked Questions." W3C. (6/27/2008) http://www.w3.org/People/Berners-Lee/FAQ.html
  • "Brief Timeline of the Internet." Webopedia Internet.com. 5/24/2007. (6/23/2008) http://www.webopedia.com/quick_ref/timeline.asp
  • Gralla, Preston. "How the Internet Works." Que. 8/1999. (6/27/2008)
  • Haynal, Russ. "Internet: The Big Picture." (6/23/2008) http://navigators.com/internet_architecture.html
  • "Internet History." Computer History Museum. 2006. (6/23/2008) http://www.computerhistory.org/internet_history/
  • Strickland, Jonathan. "How ARPANET Works." HowStuffWorks.com. (6/23/2008). https://computer.howstuffworks.com/arpanet.htm
  • Strickland, Jonathan. "How did the Internet start?" HowStuffWorks.com. (6/23/2008). https://computer.howstuffworks.com/internet-start.htm
  • "The Difference Between the Internet and the World Wide Web." Webopedia Internet.com. 2/29/2008. (6/23/2008) http://www.webopedia.com/didyouknow/internet/2002/web_vs_internet.asp
  • Tyson, Jeff. "How Internet Infrastructure Works." HowStuffWorks.com. (6/23/2008). https://computer.howstuffworks.com/internet-infrastructure.htm
  • "What is HTTP?" WhatIs.com. 7/27/2006. (6/27/2008) http://searchwindevelopment.techtarget.com/sDefinition/0,,sid8_gci214004,00.html
  • "What is World Wide Web." WhatIs.com. 3/6/2007. (6/27/2008) http://searchcrm.techtarget.com/sDefinition/0,,sid11_gci213391,00.html
  • "Who owns the Internet." Webopedia Internet.com. 2/29/2008. (6/23/2008) http://www.webopedia.com/DidYouKnow/Internet/2002/WhoOwnstheInternet.asp

Please copy/paste the following text to properly cite this HowStuffWorks.com article:

world wide web 30

National Academies Press: OpenBook

Funding a Revolution: Government Support for Computing Research (1999)

Chapter: 7 development of the internet and the world wide web, 7 development of the internet and the world wide web.

The recent growth of the Internet and the World Wide Web makes it appear that the world is witnessing the arrival of a completely new technology. In fact, the Web—now considered to be a major driver of the way society accesses and views information—is the result of numerous projects in computer networking, mostly funded by the federal government, carried out over the last 40 years. The projects produced communications protocols that define the format of network messages, prototype networks, and application programs such as browsers. This research capitalized on the ubiquity of the nation's telephone network, which provided the underlying physical infrastructure upon which the Internet was built.

This chapter traces the development of the Internet, 1 one aspect of the broader field of data networking. The chapter is not intended to be comprehensive; rather, it focuses on the federal role in both funding research and supporting the deployment of networking infrastructure. This history is divided into four distinct periods. Before 1970, individual researchers developed the underlying technologies, including queuing theory, packet switching, and routing. During the 1970s, experimental networks, notably the ARPANET, were constructed. These networks were primarily research tools, not service providers. Most were federally funded, because, with a few exceptions, industry had not yet realized the potential of the technology. During the 1980s, networks were widely deployed, initially to support scientific research. As their potential to improve personal communications and collaboration became apparent, additional academic disciplines and industry began to use the technol-

ogy. In this era, the National Science Foundation (NSF) was the major supporter of networking, primarily through the NSFNET, which evolved into the Internet. Most recently, in the early 1990s, the invention of the Web made it much easier for users to publish and access information, thereby setting off the rapid growth of the Internet. The final section of the chapter summarizes the lessons to be learned from history.

By focusing on the Internet, this chapter does not address the full scope of computer networking activities that were under way between 1960 and 1995. It specifically ignores other networking activities of a more proprietary nature. In the mid-1980s, for example, hundreds of thousands of workers at IBM were using electronic networks (such as the VNET) for worldwide e-mail and file transfers; banks were performing electronic funds transfer; Compuserve had a worldwide network; Digital Equipment Corporation (DEC) had value-added networking services; and a VNET-based academic network known as BITNET had been established. These were proprietary systems that, for the most part, owed little to academic research, and indeed were to a large extent invisible to the academic computer networking community. By the late 1980s, IBM's proprietary SNA data networking business unit already had several billions of dollars of annual revenue for networking hardware, software, and services. The success of such networks in many ways limited the interest of companies like IBM and Compuserve in the Internet. The success of the Internet can therefore, in many ways, be seen as the success of an open system and open architecture in the face of proprietary competition.

Early Steps: 1960-1970

Approximately 15 years after the first computers became operational, researchers began to realize that an interconnected network of computers could provide services that transcended the capabilities of a single system. At this time, computers were becoming increasingly powerful, and a number of scientists were beginning to consider applications that went far beyond simple numerical calculation. Perhaps the most compelling early description of these opportunities was presented by J.C.R. Licklider (1960), who argued that, within a few years, computers would become sufficiently powerful to cooperate with humans in solving scientific and technical problems. Licklider, a psychologist at the Massachusetts Institute of Technology (MIT), would begin realizing his vision when he became director of the Information Processing Techniques Office (IPTO) at the Advanced Research Projects Agency (ARPA) in 1962. Licklider remained at ARPA until 1964 (and returned for a second tour in 1974-1975), and he convinced his successors, Ivan Sutherland and Robert Taylor, of the importance of attacking difficult, long-term problems.

Taylor, who became IPTO director in 1966, worried about the duplication of expensive computing resources at the various sites with ARPA contracts. He proposed a networking experiment in which users at one site accessed computers at another site, and he co-authored, with Licklider, a paper describing both how this might be done and some of the potential consequences (Licklider and Taylor, 1968). Taylor was a psychologist, not a computer scientist, and so he recruited Larry Roberts of MIT's Lincoln Laboratory to move to ARPA and oversee the development of the new network. As a result of these efforts, ARPA became the primary supporter of projects in networking during this period.

In contrast to the NSF, which awarded grants to individual researchers, ARPA issued research contracts. The IPTO program managers, typically recruited from academia for 2-year tours, had considerable latitude in defining projects and identifying academic and industrial groups to carry them out. In many cases, they worked closely with the researchers they sponsored, providing intellectual leadership as well as financial support. A strength of the ARPA style was that it not only produced artifacts that furthered its missions but also built and trained a community of researchers. In addition to holding regular meetings of principal investigators, Taylor started the "ARPA games," meetings that brought together the graduate students involved in programs. This innovation helped build the community that would lead the expansion of the field and growth of the Internet during the 1980s.

During the 1960s, a number of researchers began to investigate the technologies that would form the basis for computer networking. Most of this early networking research concentrated on packet switching, a technique of breaking up a conversation into small, independent units, each of which carries the address of its destination and is routed through the network independently. Specialized computers at the branching points in the network can vary the route taken by packets on a moment-to-moment basis in response to network congestion or link failure.

One of the earliest pioneers of packet switching was Paul Baran of the RAND Corporation, who was interested in methods of organizing networks to withstand nuclear attack. (His research interest is the likely source of a widespread myth concerning the ARPANET's original purpose [Hafner and Lyon, 1996]). Baran proposed a richly interconnected set of network nodes, with no centralized control system—both properties of today's Internet. Similar work was under way in the United Kingdom, where Donald Davies and Roger Scantlebury of the National Physical Laboratory (NPL) coined the term "packet."

Of course, the United States already had an extensive communications network, the public switched telephone network (PSTN), in which digital switches and transmission lines were deployed as early as 1962.

But the telephone network did not figure prominently in early computer networking. Computer scientists working to interconnect their systems spoke a different language than did the engineers and scientists working in traditional voice telecommunications. They read different journals, attended different conferences, and used different terminology. Moreover, data traffic was (and is) substantially different from voice traffic. In the PSTN, a continuous connection, or circuit, is set up at the beginning of a call and maintained for the duration. Computers, on the other hand, communicate in bursts, and unless a number of "calls" can be combined on a single transmission path, line and switching capacity is wasted. Telecommunications engineers were primarily interested in improving the voice network and were skeptical of alternative technologies. As a result, although telephone lines were used to provide point-to-point communication in the ARPANET, the switching infrastructure of the PSTN was not used. According to Taylor, some Bell Laboratories engineers stated flatly in 1967 that "packet switching wouldn't work." 2

At the first Association for Computing Machinery (ACM) Symposium on Operating System Principles in 1967, Lawrence Roberts, then an IPTO program manager, presented an initial design for the packet-switched network that was to become the ARPANET (Davies et al., 1967). In addition, Roger Scantlebury presented the NPL work (Roberts, 1967), citing Baran's earlier RAND report. The reaction was positive, and Roberts issued a request for quotation (RFQ) for the construction of a four-node network.

From the more than 100 respondents to the RFQ, Roberts selected Bolt, Beranek, and Newman (BBN) of Cambridge, Massachusetts; familiar names such as IBM Corporation and Control Data Corporation chose not to bid. The contract to produce the hardware and software was issued in December 1968. The BBN group was led by Frank Heart, and many of the scientists and engineers who would make major contributions to networking in future years participated. Robert Kahn, who with Vinton Cerf would later develop the Transmission Control Protocol/Internet Protocol (TCP/IP) suite used to control the transmission of packets in the network, helped develop the network architecture. The network hardware consisted of a rugged military version of a Honeywell Corporation minicomputer that connected a site's computers to the communication lines. These interface message processors (IMPs)—each the size of a large refrigerator and painted battleship gray—were highly sought after by DARPA-sponsored researchers, who viewed possession of an IMP as evidence they had joined the inner circle of networking research.

The first ARPANET node was installed in September 1969 at Leonard Kleinrock's Network Measurement Center at the University of California at Los Angeles (UCLA). Kleinrock (1964) had published some of the

earliest theoretical work on packet switching, and so this site was an appropriate choice. The second node was installed a month later at Stanford Research Institute (SRI) in Menlo Park, California, using Douglas Engelbart's On Line System (known as NLS) as the host. SRI also operated the Network Information Center (NIC), which maintained operational and standards information for the network. Two more nodes were soon installed at the University of California at Santa Barbara, where Glen Culler and Burton Fried had developed an interactive system for mathematics education, and the University of Utah, which had one of the first computer graphics groups.

Initially, the ARPANET was primarily a vehicle for experimentation rather than a service, because the protocols for host-to-host communication were still being developed. The first such protocol, the Network Control Protocol (NCP), was completed by the Network Working Group (NWG) led by Stephen Crocker in December 1970 and remained in use until 1983, when it was replaced by TCP/IP.

Expansion of the Arpanet: 1970-1980

Initially conceived as a means of sharing expensive computing resources among ARPA research contractors, the ARPANET evolved in a number of unanticipated directions during the 1970s. Although a few experiments in resource sharing were carried out, and the Telnet protocol was developed to allow a user on one machine to log onto another machine over the network, other applications became more popular.

The first of these applications was enabled by the File Transfer Protocol (FTP), developed in 1971 by a group led by Abhay Bhushan of MIT (Bhushan, 1972). This protocol enabled a user on one system to connect to another system for the purpose of either sending or retrieving a particular file. The concept of an anonymous user was quickly added, with constrained access privileges, to allow users to connect to a system and browse the available files. Using Telnet, a user could read the remote files but could not do anything with them. With FTP, users could now move files to their own machines and work with them as local files. This capability spawned several new areas of activity, including distributed client-server computing and network-connected file systems.

Occasionally in computing, a "killer application" appears that becomes far more popular than its developers expected. When personal computers (PCs) became available in the 1980s, the spreadsheet (initially VisiCalc) was the application that accelerated the adoption of the new hardware by businesses. For the newly minted ARPANET, the killer application was electronic mail, or e-mail. The first e-mail program was developed in 1972 by Ray Tomlinson of BBN. Tomlinson had built an

earlier e-mail system for communication between users on BBN's Tenex time-sharing system, and it was a simple exercise to modify this system to work over the network. By combining the immediacy of the telephone with the precision of written communication, e-mail became an instant hit. Tomlinson's syntax ( user@domain ) remains in use today.

Telnet, FTP, and e-mail were examples of the leverage that research typically provided in early network development. As each new capability was added, the efficiency and speed with which knowledge could be disseminated improved. E-mail and FTP made it possible for geographically distributed researchers to collaborate and share results much more effectively. These programs were also among the first networking applications that were valuable not only to computer scientists, but also to scholars in other disciplines.

From Arpanet to Internet

Although the ARPANET was ARPA's largest networking effort, it was by no means the only one. The agency also supported research on terrestrial packet radio and packet satellite networks. In 1973, Robert Kahn and Vinton Cerf began to consider ways to interconnect these networks, which had quite different bandwidth, delay, and error properties than did the telephone lines of the ARPANET. The result was TCP/IP, first described in 1973 at an International Network Working Group meeting in England. Unlike NCP, which enabled the hosts of a single network to communicate, TCP/IP was designed to interconnect multiple networks to form an Internet. This protocol suite defined the packet format and a flow-control and error-recovery mechanism to allow the hosts to recover gracefully from network errors. It also specified an addressing mechanism that could support an Internet comprising up to 4 billion hosts.

The work necessary to transform TCP/IP from a concept into a useful system was performed under ARPA contract by groups at Stanford University, BBN, and University College London. Although TCP/IP has evolved over the years, it is still in use today as the Internet's basic packet transport protocol.

By 1975, the ARPANET had grown from its original four nodes to nearly 100 nodes. Around this time, two phenomena—the development of local area networks (LANs) and the integration of networking into operating systems—contributed to a rapid increase in the size of the network.

Local Area Networks

While ARPANET researchers were experimenting with dedicated telephone lines for packet transmission, researchers at the University of

Hawaii, led by Norman Abramson, were trying a different approach, also with ARPA funding. Like the ARPANET group, they wanted to provide remote access to their main computer system, but instead of a network of telephone lines, they used a shared radio network. It was shared in the sense that all stations used the same channel to reach the central station. This approach had a potential drawback: if two stations attempted to transmit at the same time, then their transmissions would interfere with each other, and neither one would be received. But such interruptions were unlikely because the data were typed on keyboards, which sent very short pulses to the computer, leaving ample time between pulses during which the channel was clear to receive keystrokes from a different user.

Abramson's system, known as Aloha, generated considerable interest in using a shared transmission medium, and several projects were initiated to build on the idea. Two of the best-known projects were the Atlantic Packet Satellite Experiment and Ethernet. The packet satellite network demonstrated that the protocols developed in Aloha for handling contention between simultaneous users, combined with more traditional reservation schemes, resulted in efficient use of the available bandwidth. However, the long latency inherent in satellite communications limited the usefulness of this approach.

Ethernet, developed by a group led by Robert Metcalfe at Xerox Corporation's Palo Alto Research Center (PARC), is one of the few examples of a networking technology that was not directly funded by the government. This experiment demonstrated that using coaxial cable as a shared medium resulted in an efficient network. Unlike the Aloha system, in which transmitters could not receive any signals, Ethernet stations could detect that collisions had occurred, stop transmitting immediately, and retry a short time later (at random). This approach improved the efficiency of the Aloha technique and made it practical for actual use. Shared-media LANs became the dominant form of computer-to-computer communication within a building or local area, although variations from IBM (Token Ring) and others also captured part of this emerging market.

Ethernet was initially used to connect a network of approximately 100 of PARC's Alto PCs, using the center's time-sharing system as a gateway to the ARPANET. Initially, many believed that the small size and limited performance of PCs would preclude their use as network hosts, but, with DARPA funding, David Clark's group at MIT, which had received several Altos from PARC, built an efficient TCP implementation for that system and, later, for the IBM PC. The proliferation of PCs connected by LANs in the 1980s dramatically increased the size of the Internet.

Integrated Networking

Until the 1970s, academic computer science research groups used a variety of computers and operating systems, many of them constructed by the researchers themselves. Most were time-sharing systems that supported a number of simultaneous users. By 1970, many groups had settled on the Digital Equipment Corporation (DEC) PDP-10 computer and the Tenex operating system developed at BBN. This standardization enabled researchers at different sites to share software, including networking software.

By the late 1970s, the Unix operating system, originally developed at Bell Labs, had become the system of choice for researchers, because it ran on DEC's inexpensive (relative to other systems) VAX line of computers. During the late 1970s and early 1980s, an ARPA-funded project at the University of California at Berkeley (UC-Berkeley) produced a version of Unix (the Berkeley System Distribution, or BSD) that included tightly integrated networking capabilities. The BSD was rapidly adopted by the research community because the availability of source code made it a useful experimental tool. In addition, it ran on both VAX machines and the personal workstations provided by the fledgling Sun Microsystems, Inc., several of whose founders came from the Berkeley group. The TCP/IP suite was now available on most of the computing platforms used by the research community.

Standards and Management

Unlike the various telecommunications networks, the Internet has no owner. It is a federation of commercial service providers, local educational networks, and private corporate networks, exchanging packets using TCP/IP and other, more specialized protocols. To become part of the Internet, a user need only connect a computer to a port on a service provider's router, obtain an IP address, and begin communicating. To add an entire network to the Internet is a bit trickier, but not extraordinarily so, as demonstrated by the tens of thousands of networks with tens of millions of hosts that constitute the Internet today.

The primary technical problem in the Internet is the standardization of its protocols. Today, this is accomplished by the Internet Engineering Task Force (IETF), a voluntary group interested in maintaining and expanding the scope of the Internet. Although this group has undergone many changes in name and makeup over the years, it traces its roots directly to Stephen Crocker's NWG, which defined the first ARPANET protocol in 1969. The NWG defined the system of requests for comments (RFCs) that are still used to specify protocols and discuss other engineer-

ing issues. Today's RFCs are still formatted as they were in 1969, eschewing the decorative fonts and styles that pervade today's Web.

Joining the IETF is a simple matter of asking to be placed on its mailing list, attending thrice-yearly meetings, and participating in the work. This grassroots group is far less formal than organizations such as the International Telecommunications Union, which defines telephony standards through the work of members who are essentially representatives of various governments. The open approach to Internet standards reflects the academic roots of the network.

Closing the Decade

The 1970s were a time of intensive research in networking. Much of the technology used today was developed during this period. Several networks other than ARPANET were assembled, primarily for use by computer scientists in support of their own research. Most of the work was funded by ARPA, although the NSF provided educational support for many researchers and was beginning to consider establishing a large-scale academic network.

During this period, ARPA pursued high-risk research with the potential for high payoffs. Its work was largely ignored by AT&T, and the major computer companies, notably IBM and DEC, began to offer proprietary networking solutions that competed with, rather than applied, the ARPA-developed technologies. 3 Yet the technologies developed under ARPA contract ultimately resulted in today's Internet. It is debatable whether a more risk-averse organization lacking the hands-on program management style of ARPA could have produced the same result.

Operation of the ARPANET was transferred to the Defense Communication Agency in 1975. By the end of the decade, the ARPANET had matured sufficiently to provide services. It remained in operation until 1989, when it was superseded by subsequent networks. The stage was now set for the Internet, which was first used by scientists, then by academics in many disciplines, and finally by the world at large.

The NSFNET Years: 1980-1990

During the late 1970s, several networks were constructed to serve the needs of particular research communities. These networks—typically funded by the federal agency that was the primary supporter of the research area—included MFENet, which the Department of Energy established to give its magnetic fusion energy researchers access to supercomputers, and NASA's Space Physics Analysis Network (SPAN). The NSF began supporting network infrastructure with the establishment

of CSNET, which was intended to link university computer science departments with the ARPANET. The CSNET had one notable property that the ARPANET lacked: it was open to all computer science researchers, whereas only ARPA contractors could use the ARPANET. An NSF grant to plan the CSNET was issued to Larry Landweber at the University of Wisconsin in 1980.

The CSNET was used throughout the 1980s, but as it and other regional networks began to demonstrate their usefulness, the NSF launched a much more ambitious effort, the NSFNET. From the start, the NSFNET was designed to be a network of networks—an ''internet''—with a high-speed backbone connecting NSF's five supercomputer centers and the National Center for Atmospheric Research. To oversee the new network, the NSF hired Dennis Jennings from Trinity College, Dublin. In the early 1980s, Jennings had been responsible for the Irish Higher Education Authority network (HEANet), and so he was well-qualified for the task. One of Jennings' first decisions was to select TCP/IP as the primary protocol suite for the NFSNET.

Because the NSFNET was to be an internet (the beginning of today's Internet), specialized computers called routers were needed to pass traffic between networks at the points where the networks met. Today, routers are the primary products of multibillion-dollar companies (e.g., Cisco Systems Incorporated, Bay Networks), but in 1985, few commercial products were available. The NSF chose the "Fuzzball" router designed by David Mills at the University of Delaware (Mills, 1988). Working with ARPA support, Mills improved the protocols used by the routers to communicate the network topology among themselves, a critical function in a large-scale network.

Another technology required for the rapidly growing Internet was the Domain Name Service (DNS). Developed by Paul Mockapetris at the University of Southern California's Information Sciences Institute, the DNS provides for hierarchical naming of hosts. An administrative entity, such as a university department, can assign host names as it wishes. It also has a domain name, issued by the higher-level authority of which it is a part. (Thus, a host named xyz in the computer science department at UC-Berkeley would be named xyz.cs.berkeley.edu. ) Servers located throughout the Internet provide translation between the host names used by human users and the IP addresses used by the Internet protocols. The name-distribution scheme has allowed the Internet to grow much more rapidly than would be possible with centralized administration.

Jennings left the NSF in 1986. He was succeeded by Stephen Wolff, who oversaw the deployment and growth of the NSFNET. During Wolff's tenure, the speed of the backbone, originally 56 kilobits per second, was increased 1,000-fold, and a large number of academic and regional net-

works were connected to the NSFNET. The NSF also began to expand the reach of the NSFNET beyond its supercomputing centers through its Connections program, which targeted the research and education community. In response to the Connections solicitation, the NSF received innovative proposals from what would become two of the major regional networks: SURANET and NYSERNET. These groups proposed to develop regional networks with a single connection to the NSFNET, instead of connecting each institution independently.

Hence, the NSFNET evolved into a three-tiered structure in which individual institutions connected to regional networks that were, in turn, connected to the backbone of the NSFNET. The NSF agreed to provide seed funding for connecting regional networks to the NSFNET, with the expectation that, as a critical mass was reached, the private sector would take over the management and operating costs of the Internet. This decision helped guide the Internet toward self-sufficiency and eventual commercialization (Computer Science and Telecommunications Board, 1994).

As the NSFNET expanded, opportunities for privatization grew. Wolff saw that commercial interests had to participate and provide financial support if the network were to continue to expand and evolve into a large, single internet. The NSF had already (in 1987) contracted with Merit Computer Network Incorporated at the University of Michigan to manage the backbone. Merit later formed a consortium with IBM and MCI Communications Corporation called Advanced Network and Services (ANS) to oversee upgrades to the NSFNET. Instead of reworking the existing backbone, ANS added a new, privately owned backbone for commercial services in 1991. 4

Emergence of the Web: 1990 to the Present

By the early 1990s, the Internet was international in scope, and its operation had largely been transferred from the NSF to commercial providers. Public access to the Internet expanded rapidly thanks to the ubiquitous nature of the analog telephone network and the availability of modems for connecting computers to this network. Digital transmission became possible throughout the telephone network with the deployment of optical fiber, and the telephone companies leased their broadband digital facilities for connecting routers and regional networks to the developers of the computer network. In April 1995, all commercialization restrictions on the Internet were lifted. Although still primarily used by academics and businesses, the Internet was growing, with the number of hosts reaching 250,000. Then the invention of the Web catapulted the Internet to mass popularity almost overnight.

The idea for the Web was simple: provide a common format for

documents stored on server computers, and give each document a unique name that can be used by a browser program to locate and retrieve the document. Because the unique names (called universal resource locators, or URLs) are long, including the DNS name of the host on which they are stored, URLs would be represented as shorter hypertext links in other documents. When the user of a browser clicks a mouse on a link, the browser retrieves and displays the document named by the URL.

This idea was implemented by Timothy Berners-Lee and Robert Cailliau at CERN, the high-energy physics laboratory in Geneva, Switzerland, funded by the governments of participating European nations. Berners-Lee and Cailliau proposed to develop a system of links between different sources of information. Certain parts of a file would be made into nodes, which, when called up, would link the user to other, related files. The pair devised a document format called HYpertext Markup Language (HTML), a variant of the Standard Generalized Markup Language used in the publishing industry since the 1950s. It was released at CERN in May 1991. In July 1992, a new Internet protocol, the Hypertext Transfer Protocol (HTTP), was introduced to improve the efficiency of document retrieval. Although the Web was originally intended to improve communications within the physics community at CERN, it—like e-mail 20 years earlier—rapidly became the new killer application for the Internet.

The idea of hypertext was not new. One of the first demonstrations of a hypertext system, in which a user could click a mouse on a highlighted word in a document and immediately access a different part of the document (or, in fact, another document entirely), occurred at the 1967 Fall Joint Computer Conference in San Francisco. At this conference, Douglas Engelbart of SRI gave a stunning demonstration of his NLS (Engelbart, 1986), which provided many of the capabilities of today's Web browsers, albeit limited to a single computer. Engelbart's Augment project was supported by funding from NASA and ARPA. Engelbart was awarded the Association for Computing Machinery's 1997 A. M. Turing Award for this work. Although it never became commercially successful, the mouse-driven user interface inspired researchers at Xerox PARC, who were developing personal computing technology.

Widespread use of the Web, which now accounts for the largest volume of Internet traffic, was accelerated by the development in 1993 of the Mosaic graphical browser. This innovation, by Marc Andreessen at the NSF-funded National Center for Supercomputer Applications, enabled the use of hyperlinks to video, audio, and graphics, as well as text. More important, it provided an effective interface that allowed users to point-and-click on a menu or fill in a blank to search for information.

The development of the Internet and the World Wide Web has had a tremendous impact on the U.S. economy and society more broadly. By

January 1998, almost 30 million host computers were connected to the Internet (Zakon, 1998), and more than 58 million users in the United States and Canada were estimated to be online (Nielsen Media Research, 1997). Numerous companies now sell Internet products worth billions of dollars. Cisco Systems, a leader in network routing technology, for example, reported sales of $8.5 billion in 1998. Netscape Communications Corporation, which commercialized the Mosaic browser, had sales exceeding $530 million in 1997. 5 Microsoft Corporation also entered the market for Web browsers and now competes head-to-head with Netscape. A multitude of other companies offer hardware and software for Internet based systems.

The Internet has also paved the way for a host of services. Companies like Yahoo! and InfoSeek provide portals to the Internet and have attracted considerable attention from Wall Street investors. Other companies, like Amazon.com and Barnes & Noble, have established online stores. Amazon had online sales of almost $150 million for books in 1997. 6 Electronic commerce, more broadly, is taking hold in many types of organizations, from PC manufacturers to retailers to travel agencies. Although estimates of the value of these services vary widely, they all reflect a growing sector of the economy that is wholly dependent on the Internet. Internet retailing could reach $7 billion by the year 2000, and online sales of travel services are expected to approach $8 billion around the turn of the century. Forrester Research estimates that businesses will buy and sell $327 billion worth of goods over the Internet by the year 2002 (Blane, 1997).

The Web has been likened to the world's largest library—with the books piled in the middle of the floor. Search engines, which are programs that follow the Web's hypertext links and index the material they discover, have improved the organization somewhat but are difficult to use, frequently deluging the user with irrelevant information. Although developments in computing and networking over the last 40 years have realized some of the potential described by visionaries such as Licklider and Engelbart, the field continues to offer many opportunities for innovation.

Lessons from History

The development of the Internet demonstrates that federal support for research, applied at the right place and right time, can be extremely effective. DARPA's support gave visibility to the work of individual researchers on packet switching and resulted in the development of the first large-scale packet-switched network. Continued support for experimentation led to the development of networking protocols and applications, such as e-mail, that were used on the ARPANET and, subsequently, the Internet.

By bringing together a diverse mix of researchers from different institutions, such federal programs helped the Internet gain widespread acceptance and established it as a dominant mode of internetworking. Government programs such as ARPANET and NSFNET created a large enough base of users to make the Internet more attractive in many applications than proprietary networking systems being offered by a number of vendors. Though a number of companies continue to sell proprietary systems for wide area networking, some of which are based on packet-switched technology, these systems have not achieved the ubiquity of the Internet and are used mainly within private industry.

Research in packet switching evolved in unexpected directions and had unanticipated consequences. It was originally pursued to make more-efficient use of limited computing capabilities and later seen as a means of linking the research and education communities. The most notable result, however, was the Internet, which has dramatically improved communication across society, changing the way people work, play, and shop. Although DARPA and the NSF were successful in creating an expansive packet-switched network to facilitate communication among researchers, it took the invention of the Web and its browsers to make the Internet more broadly accessible and useful to society.

The widespread adoption of Internet technology has created a number of new companies in industries that did not exist 20 years ago, and most companies that did exist 20 years ago are incorporating Internet technology into their business operations. Companies such as Cisco Systems, Netscape Communications, Yahoo!, and Amazon.com are built on Internet technologies and their applications and generate billions of dollars annually in combined sales revenues. Electronic commerce is also maturing into an established means of conducting business.

The complementary missions and operating styles of federal agencies are important to the development and implementation of new technologies. Whereas DARPA supported early research on packet switching and development of the ARPANET, it was not prepared to support an operational network, nor did it expand its network beyond DARPA-supported research institutions. With its charter to support research and education, the NSF both supported an operational network and greatly expanded its reach, effectively building the infrastructure for the Internet.

The past 50 years have witnessed a revolution in computing and related communications technologies. The contributions of industry and university researchers to this revolution are manifest; less widely recognized is the major role the federal government played in launching the computing revolution and sustaining its momentum. Funding a Revolution examines the history of computing since World War II to elucidate the federal government's role in funding computing research, supporting the education of computer scientists and engineers, and equipping university research labs. It reviews the economic rationale for government support of research, characterizes federal support for computing research, and summarizes key historical advances in which government-sponsored research played an important role.

Funding a Revolution contains a series of case studies in relational databases, the Internet, theoretical computer science, artificial intelligence, and virtual reality that demonstrate the complex interactions among government, universities, and industry that have driven the field. It offers a series of lessons that identify factors contributing to the success of the nation's computing enterprise and the government's role within it.

READ FREE ONLINE

Welcome to OpenBook!

You're looking at OpenBook, NAP.edu's online reading room since 1999. Based on feedback from you, our users, we've made some improvements that make it easier than ever to read thousands of publications on our website.

Do you want to take a quick tour of the OpenBook's features?

Show this book's table of contents , where you can jump to any chapter by name.

...or use these buttons to go back to the previous chapter or skip to the next one.

Jump up to the previous page or down to the next one. Also, you can type in a page number and press Enter to go directly to that page in the book.

Switch between the Original Pages , where you can read the report as it appeared in print, and Text Pages for the web version, where you can highlight and search the text.

To search the entire text of this book, type in your search term here and press Enter .

Share a link to this book page on your preferred social network or via email.

View our suggested citation for this chapter.

Ready to take your reading offline? Click here to buy this book in print or download it as a free PDF, if available.

Get Email Updates

Do you enjoy reading reports from the Academies online for free ? Sign up for email notifications and we'll let you know about new publications in your areas of interest when they're released.

World Wide Web

Assignment #1, assignment #2, search engines, assignments, internet search, foreign language search engines.

  • Full IELTS Practice Tests
  • Practice Tests

The World Wide Web from its origins

  • View Solution

Solution for: The World Wide Web from its origins

Answer table.

 Found a mistake? Let us know!

 Share this Practice Test

Exam Review

Highlight

Science inspired the World Wide Web, and the Web has responded by changing science.

'Information Management: A Proposal' . That was the bland title of a document written in March 1989 by a then little-known computer scientist called Tim Berners-Lee, who was working at CERN, Europe’s particle physics laboratory, near Geneva. His proposal, modestly called the World Wide Web, has achieved far more than anyone expected at the time.

In fact, the Web was invented to deal with a specific problem. In the late 1980s, CERN was planning one of the most ambitious scientific projects ever , the Large Hadron Collider*, or LHC. As the first few lines of the original proposal put it, 'Many of the discussions of the future at CERN and the LHC end with the question "Yes, but how will we ever keep track of such a large project?"  This proposal provides an answer to such questions.

The Web, as everyone now knows, has many more uses than the original idea of linking electronic documents about particle physics in laboratories around the world. But among all the changes it has brought about, from personal social networks to political campaigning, it has also transformed the business of doing science itself, as the man who invented it hoped it would.

It allows journals to be published online and links to be made from one paper to another. It also permits professional scientists to recruit thousands of amateurs to give them a hand . One project of this type, called GalaxyZoo, used these unpaid workers to classify one million images of galaxies into various types (spiral, elliptical and irregular). This project, which was intended to help astronomers understand how galaxies evolve, was so successful that a successor has now been launched, to classify the brightest quarter of a million of them in finer detail. People working for a more modest project called Herbaria@home examine scanned images of handwritten notes about old plants stored in British museums. This will allow them to track the changes in the distribution of species in response to climate change.

Another new scientific application of the Web is to use it as an experimental laboratory. It is allowing social scientists, in particular, to do things that were previously impossible. In one project, scientists made observations about the sizes of human social networks using data from Facebook. A second investigation of these networks, produced by Bernardo Huberman of HP Labs, Hewlett-Packard's research arm in Pato Alto, California, looked at Twitter, a social networking website that allows people to post short messages to long lists of friends.

At first glance, the networks seemed enormous - the 300,000 Twitterers sampled had 80 friends each, on average (those on Facebook had 120), but some listed up to 1,000. Closer statistical inspection, however, revealed that the majority of the messages were directed at a few specific friends. This showed that an individual's active social network is far smaller than his 'clan'. Dr Huberman has also helped uncover several laws of web surfing , including the number of times an average person will go from web page to web page on a given site before giving up, and the details of the 'winner takes all' phenomenon, whereby a few sites on a given subject attract most of the attention , and the rest get very little.

Scientists have been good at using the Web to carry out research. However, they have not been so effective at employing the latest web-based social-networking tools to open up scientific discussion and encourage more effective collaboration.  Journalists are now used to having their articles commented on by dozens of readers. Indeed, many bloggers develop and refine their essays as a result of these comments.

Yet although people have tried to have scientific research reviewed in the same way, most researchers only accept reviews from a few anonymous experts. When Nature , one of the world's most respected scientific journals, experimented with open peer review in 2006, the results were disappointing. Only 5% of the authors it spoke to agreed to have their article posted for review on the Web - and their instinct turned out to be right, because almost half of the papers attracted no comments. Michael Nielsen, an expert on quantum computers, belongs to a new wave of scientist who want to change this. He thinks the reason for the lack of comments is that potential reviewers lack incentive.

adapted from The Economist

* The Large Hadron Collider (LHC) is the world's largest particle accelerator and collides particle beams. It provides information on fundamental questions of physics.

Questions 1-6

Do the following statements agree with the information given in the reading passage?

Write              

TRUE            if the statement agrees with the  information

FALSE          if the statement contradicts the  information

NOT GIVEN if there is no information on this

1 TRUE FALSE NOT GIVEN     Tim Berners-Lee was famous for his research in physics before he invented the World Wide Web. Answer: FALSE      Locate

2 TRUE FALSE NOT GIVEN     The original intention of the Web was to help manage one extremely complex project. Answer: TRUE      Locate

3 TRUE FALSE NOT GIVEN     Tim Berners-Lee has also been active in politics. Answer: NOT GIVEN

4 TRUE FALSE NOT GIVEN     The Web has allowed professional and amateur scientists to work together. Answer: TRUE      Locate

5 TRUE FALSE NOT GIVEN     The second galaxy project aims to examine more galaxies than the first. Answer: FALSE      Locate

6 TRUE FALSE NOT GIVEN     Herbaria@home’s work will help to reduce the effects of climate change. Answer: NOT GIVEN

Questions 7-10

Complete the notes below.

Choose NO MORE THAN TWO WORDS from the passage for each answer.

Social networks and Internet use

Web used by Social scientists (including  Dr Huberman) to investigate the 7   of social networks. Answer: sizes      Locate

Most 8   intended for limited number  of people - not everyone on list. Answer: messages      Locate

Dr Huberman has also investigated:

  •     9   to discover how long people will sp end on a particular website, Answer: web surfing      Locate
  •     why a small number of sites get much more  10 than others on same subject. Answer: attention      Locate

Questions 11-13

Answer the questions below.

Choose NO MORE THAN TWO WORDS from the  passage for each answer.

11 Whose writing improves as a result of feedback received from readers? Answer: bloggers

12 What type of writing is not reviewed extensively on the Web? Answer: scientific research      Locate

13  Which publication invited authors to publish their articles on the World Wide Web? Answer: Nature      Locate

Other Tests

  • 4 - Multiple Choice
  • 6 - Matching Headings
  • 3 - Matching Information

Deforestation in the 21st century

  • Nature & Environment
  • 0 unanswered
  • 10 - Matching Headings
  • 4 - Summary, form completion

Green virtues of green sand

  • 6 - YES-NO-NOT GIVEN
  • 2 - Sentence Completion
  • 6 - Summary, form completion

Worldly Wealth

  • 8 - Sentence Completion

A neuroscientist reveals how to think differently

  • 3 - Multiple Choice
  • 7 - YES-NO-NOT GIVEN
  • 3 - Sentence Completion

Zoo conservation programmes

  • 5 - YES-NO-NOT GIVEN
  • 5 - Matching Information
  • 4 - Sentence Completion

Museums of fine art and their public

  • Leisure & Entertainment

Found a mistake? Let us know!

Please descibe the mistake as details as possible along with your expected correction, leave your email so we can contact with you when needed.

Describe what is wrong with the practice test:

Please enter description

Enter your name:

Enter your email address:

Please enter a valid email

assignment on world wide web

Internet vs. World Wide Web

The World Wide Web (WWW) is one set of software services running on the Internet. The Internet itself is a global, interconnected network of computing devices. This network supports a wide variety of interactions and communications between its devices. The World Wide Web is a subset of these interactions and supports websites and URIs .

Comparison chart

The Internet is actually a huge network that is accessible to everyone & everywhere across the world. The network is composed of sub-networks comprising of a number of computers that are enabled to transmit data in packets. The internet is governed by a set of rules, laws & regulations, collectively known as the Internet Protocol (IP). The sub-networks may range from defense networks to academic networks to commercial networks to individual PCs. Internet, essentially provides information & services in the form of E-Mail, chat & file transfers. It also provides access to the World Wide Web & other interlinked web pages.

The Internet & the World Wide Web (the Web), though used interchangeably, are not synonymous. Internet is the hardware part - it is a collection of computer networks connected through either copper wires, fiber-optic cables or wireless connections whereas, the World Wide Web can be termed as the software part – it is a collection of web pages connected through hyperlinks and URLs . In short, the World Wide Web is one of the services provided by the Internet. Other services over the Internet include e-mail, chat and file transfer services. All of these services can be provided to consumers for use by businesses or government or by individuals creating their own networks or platforms.

Another method to differentiate between both is using the Protocol Suite – a collection of laws & regulations that govern the Internet. While internet is governed by the Internet Protocol – specifically dealing with data as whole and their transmission in packets , the World Wide Web is governed by the Hyper Text Transfer Protocol (HTTP) that deals with the linking of files, documents and other resources of the World Wide Web.

The Advanced Research Projects Agency (ARPA) created by the US in 1958 as a reply to the USSR’s launching of the Sputnik , led to creation of a department called the Information Processing Technology Office (IPTO) which started the Semi Automatic Ground Environment (SAGE) that linked all the radar systems of US together. With tremendous research happening across the world, the University of California in Los Angeles (UCLA) got the ARPANET , a smaller version of the Internet in 1969. Since then Internet has taken huge strides in terms of technology and connectivity to reach its current position. In 1978, the International Packet Switched Service (IPSS) was created in Europe by the British Post Office in collaboration with Tymnet & Western Union International and this network slowly spread its wings to the US and Australia. In 1983, the first Wide Area Network (WAN) was created by the National Science Foundation (NSF) of the US called the NSFnet. All these sub-networks merged together post 1985 with new definitions of the Transfer Control Protocols of the Internet Protocol ( TCP /IP) for optimization of resources.

The Web was invented by Sir Tim Berners Lee. In March 1989, Tim Berners-Lee wrote a proposal that described the Web as an elaborate information management system. With help from Robert Cailliau, he published a more formal proposal for the World Wide Web on November 12, 1990. By Christmas 1990, Berners-Lee had built all the tools necessary for a working Web: the first web browser (which was a web editor as well), the first web server , and the first Web pages which described the project itself. On August 6, 1991, he posted a short summary of the World Wide Web project on the alt.hypertext newsgroup. This date also marked the debut of the Web as a publicly available service on the Internet.

Berners-Lee's breakthrough was to marry hypertext to the Internet . In his book Weaving The Web , he explains that he had repeatedly suggested that a marriage between the two technologies was possible to members of both technical communities, but when no one took up his invitation, he finally tackled the project himself. In the process, he developed a system of globally unique identifiers for resources on the Web and elsewhere: the Uniform Resource Identifier.

The World Wide Web had a number of differences from other hypertext systems that were then available. The Web required only unidirectional links rather than bidirectional ones. This made it possible for someone to link to another resource without action by the owner of that resource. It also significantly reduced the difficulty of implementing web servers and browsers (in comparison to earlier systems), but in turn presented the chronic problem of link rot . Unlike predecessors such as HyperCard , the World Wide Web was non-proprietary, making it possible to develop servers and clients independently and to add extensions without licensing restrictions.

For more details see The History of the Internet and The History of the World Wide Web .

Internet of Things

In recent years, the phrase Internet of Things—or IoT—has been used to denote a subset of the Internet that connects physical devices, such as home appliances, vehicles, industrial sensors. Historically the devices connected to the Internet have been computers, cell phones and tablets. With the Internet of Things, other devices like refrigerators, HVAC systems, light bulbs, cars, thermostats, video cameras, and locks can also connect to the Internet. This allows better monitoring and more control of the physical world through the Internet.

About the Author

Nick Jasuja

Related Comparisons

Comcast vs FiOS

Share this comparison via:

If you read this far, you should follow us:

"Internet vs World Wide Web." Diffen.com. Diffen LLC, n.d. Web. 23 Oct 2024. < >

Comments: Internet vs World Wide Web

Anonymous comments (2).

September 12, 2012, 3:28pm This was very educating once you really get to reading these paragraphs. — 170.✗.✗.19
May 8, 2014, 9:23am Nice ans — 101.✗.✗.158
  • Comcast vs FiOS
  • Flickr vs Picasa
  • Google vs Yahoo
  • Cat5 vs Cat5e
  • WikiLeaks vs Wikipedia
  • Cable vs DSL
  • Modem vs Router

Edit or create new comparisons in your area of expertise.

Stay connected

© All rights reserved.

  • Create A Quiz
  • Relationship
  • Personality
  • Harry Potter
  • Online Exam
  • Entertainment
  • Featured Clients
  • Case Studies
  • Testimonials
  • Integrations
  • Suggestion Box
  • Ask A Question
  • Training Maker
  • Survey Maker
  • Brain Games
  • ProProfs.com

World Wide Web Quizzes, Questions & Answers

Top trending quizzes.

Radio Button

Popular Topics

Internet Quizzes, Internet Trivia, Internet Questions

Recent Quizzes

Assignment on World Wide Web

  Web  Site:

A  web  site  is  a  collection  of  web  pages, images, videos  or  other  digital  assets  that  is  hosted  on  one  or  several  web  server(s), usually  accessible  via  the  Internet, cell  phone  or  a  LAN. A  web  page  is  a  document,  typically  written  in  HTML, that  is  almost  always  accessible  via  HTTP, a  protocol  that  transfers  information  from  the  web  server  to  display  in  the  user’s  web  browser. All  publicity  accessible  websites  are  seen  collectively  as  constituting  the  “World  Wide  Web”.

The  pages  of  websites  can  usually  be  accessed  from  a  common  root  URL  called  the  homepage, and  usually  reside  on  the  same  physical  server. The  URLs  of  the  pages  organize  them  into  a  hierarchy, although  the  hyperlinks  between  them  control  how  the  reader  perceives  the  overall  structure  and  how  the  traffic  flows  between  the  different  parts  of  the  sites. Some  websites  require  a  subscription  to  access  some  or  all  of  their  content. Examples  of  subscription  sites  include  many  business  sites, parts  of  many  news  sites,  academic  journal  sites, gaming  sites, message  boards, Web-based  e-mail, services, social  networking  website  and  sites  providing  real-time  stock  market  data.

Organized  by  the  function, a  website  may  be-

          »  A  personal  website

          »  A  commercial  website

          »  A  government  website

          »  A  non-profit  organization  website

It  could  be  the  work  of  an  individual, a  business  or  other  organization  and  is  typically  dedicated  to  some  particular  topic  or  purpose.

Web  The  New  Arena:

Life  was  just  going  on  fine, when  along  came  the  Internet  and  just  about  everything  changed. You  just  have  to  look  back  over  the  past  five  year  or  six  and  think  of  different  things  were  before  that. Today, everyone’s  young, smart  and  online. Oh  yes, we  are  well  into  the  Net  Age. Whether  you  are  working  in  a  high-tech  corporation  or  setting  up  your  home  office, trying  to  learn  or  two  at  college  or  having  a  whale  of  a  time  at  school, life’s  on  the  Internet .

 Content  Is  The  King:

No  matter  how  great  a  site  looks, no  amount  of  design  ever  makes  up  for  poor  content. This  is  a  fact  that  many  web  author  lose  sight  of. That’s  why  we  have  so  many  sites  around  us  that  offer  the  visitor  the  same  old  thing – a  bit  of  this  a  bit  of  that. A  really  good  site  must  have  solid  unique  content. That’s  why  as  experts  recommend  we  started  with  strategy  and  purpose  first – no  with  design. First  off  you  must  quite  clear  of  the  purpose  of  your  site. This  holds  true  for  any  type  of  web  sites, whether  it’s  a  personal  web  sites, a  small  business  set  up, a  hobbyist’s  page  or  e- commerce  or  anything  else. A  web  site  without  purpose  just  takes  space  and  please  no  one  but  its  own  author. So  unless  you  are  just  using  your  site  for  storage, start  with  putting  down  your  purpose, your  objectives, and  message.

Types  Of  Web  Site:

There  are  many  variety  of  web  sites, each  specializing  in  a  particular  type  of  content  or  use  and  they  may  be  arbitrarily  classified  in  any  number  of  ways. A  few  such  classifications  might  include:-

Affiliated  Sites:  Enabled  portal  that  renders  not  only  its  custom  CMS  but  also  syndicated  content  from  other  content  providers  for  an  agreed  fee. There  are  usually  three  relationship  tiers. Affiliate  Agencies ( e.g.  Commission  Junction),  Advertisers ( e.g.  Ebay)  and  consumer ( e.g.  Yahoo).

Archive  Site:  Used  to  preserve  valuable  electronic  content  threatened  with  extinction. Two  examples  are – Internet  Archive  which  since  1996  has  preserved  billions  of  old ( and  new )  web  pages  and  Google  Groups  which  in  early  2005  was  archiving  over  845,000,000  messages  posted  to  Usenet  news / discussion  groups.

Blog  Site:  Sites  generally  used  to  post  online  diaries  which  may  include  discussion  forums ( e.g.  blogger, xanga).

Content  Site:  Sites  whose  business  is  the  creation  and  distribution  of  original  content ( e.g.  Slant,  About.com).

Corporate  Site:  Used  to  provide  background  information  about  a  business, organization  or  service.

E-Commerce  Site:  For  purchasing  goods  such  as  Amazon.com.

Community  Site:  A  site  where  persons  with  similar  interests  communicate  with  each  other  usually  by  chat  or  message  boards  such  as  MySpace.

Database  Site:  A  site  whose  main  use  is  the  search  and  display  of  a  specific  database’s  content  such  as  the  Internet  Movie  Database  or  the  Political  Graveyard.

Development  Site:  A  site  whose  purpose  is  to  provide  information  and  resources  related  to  software  development, web  design  and  the  like.

Directory  Site:  A  site  that  contains  varied  contents  which  are  divided  into  categories  and  subcategories  such  as  Yahoo!  directory, Google  directory  and  open  directory  project.

Download  Site:  Strictly  used  for  downloading  electronic  content  such  as  software, game  demos  or  computer  wall  paper.

Employment  Site:   Allows  employers  to  post  job  requirements  for  a  position  or  positions  and  prospective  employees  to  fill  an  application.

Erotica  Websites:  Shows  sexual  videos  and  images.

Fan  Site:   A  web  site  created  and  maintained  by  fans  of  and  for  a  particular  celebrity  as  opposed  to  a  web  site  created, maintained  and  controlled  by  a  celebrity  through  their  own  paid  webmaster. May  also  be  known  as  a  Shine  in  the  case  of  certain  subjects  such  as  anime  and  manga  characters.

Game  Site:  A  site  that  is  itself  a  game  or  “playground”  where  many  people  come  to  play  such  as  MSN  Games,  POGO.com  and  Newgrounds.com.

Gripe  Site:  A  site  devoted  to  the  critique  of  a person, place, corporation, government  or  institution.

Humor  Site:  Satirizes, parodies  or  otherwise  exists  solely  to  amuse.

Information  Site:  Contains  content  that  is  intended  to  inform  visitors  but  not  necessarily  for  commercial  purposes  such  as: RateMyProfessors.com, free  internet  lexicon  and  encyclopedia. Most  government, educational  and  non-profit  institutions  have  an  informational  site.

Java  Applet  Site:  Contains  software  to  run  over  the  web  as  a  web  application.

Mirror ( Computing )  Site:  A  complete  reproduction  of  a  website.

News  Site:  Similar  to  an  information  site  but  dedicated  to  dispensing  news  and  commentary.

Personal  Homepage:  Run  by  an  individual  or  a  small  group ( such  as  a  family )  that  contains  information  or  any  content  that  the  individual  wishes  to  include.

Political  Site:  A  site  on  which  people  may  voice  political  views.

Pornography (  porn )  Site:  Site  that  shows  pornographic  images  &  videos.

Rating  Site:  Site  on  which  people  can  praise  or  disparage  what  is  featured.

Review  Site:  Site  on  which  people  can  post  reviews  for  product  or  service.

Search  Engine  Site:  A  site  that  provides  general  information  and  is  intended  as  a  gateway  or  lookup  for  other  sites. A  pure  example  is  Google  and  the  most  widely  known  extended  type  is  Yahoo! .

Shock  Site:   Includes  images  or  other  material  that  is  intended  to  be  offensive  to  most  viewers ( e.g.  Rotten.com ).

Phish  Site:   Website  created  to  fraudulently  acquire  sensitive  information  such  as  passwords  and  credit  card  details  by  masquerading  as  a  trustworthy  person  or  business  ( such  as  social  security  administration,  paypal )  in  an  electronic  communication.

Warez:   A  site  filled  with  illegal  downloads.

Web  Portal:   A  site  that  provides  a  starting  point  or  a  gateway  to  other  resources  on  the  internet  or  an  intranet.

Wiki  Site:   A  site  which  users  collaboratively  edit ( such  as  Wikipedia ).

World  Wide  Web:

The  letters  “www”  are  commonly  found  at  the  beginning  of  Web  addresses  because  of  the  long-standing  practice  if  naming  Internet  hosts ( servers )  according  to  the  services  they  provide. So  for  example, the  host  name  for a  Web  server  is  often  “www”  for  an  FTP  server, “ftp”;  and  for  a  USENET  news  server,  “news”   or   “ nntp ”  ( after  the news  protocol  NNTP). These  host  names  appear  as  DNS  subdomain  names  as  in  www.example.com. This use  of  such  prefix  is  not  require  by  any  technical  standard;  indeed,  the web  server  was  at  “ nxoc01.cern.ch”, [ 15 ] and even  today  many  web  sites  exist  without  a  “ www ”  prefix  has  no  meaning  in  the  way  the  main  web  site  is  shown.  The  “ www ”  prefix  is  simply  one  choice  for  a  web  site’s subdomain  name. Some web  browsers  will  automatically  try  adding  “ www ” to  the  beginning  and  possibly “ .com ”  to  the  end  of  typed  URLs if  no  host  is  found  without  them. Internet  Explorer, Mozilla  Firefox,  Safari  and opera  will  also  prefix  “ http://www”  and  append “ .com ” to  the  address  bar  contents  if  the  Control  and  Enter  keys  are  pressed  simulteniously. For  example, entering “ example ” in  the  address  bar  and  then  press  either  just  Enter  or  Control+Enter  will  usually  resolve  to  “ http://www.example.com ”  depending  on  the  exact  browser  version  and  its  settings.

The  World  Wide  Web ( commonly  shortened  to  the  web )  is  a  system  of  interlinked,  hypertext  documents  accessed   via  the  Internet. With  a  web  browser,  a  user  views  web  pages  that  contain  text,  images,  videos  and  other  multimedia  and  navigates  between  them  using  hyperlinks. The  World  Wide  Web  was  created  in  1989  by  Sir  Tim  Berners  Lee,  working  at  CERN  in  Geneva, Switzerland. Since  then,  Berners  Lee  has  played  an  active  role  in  guiding  the  development  of  web  standards ( such  as  the  markup  languages  in  which  web  pages  are  composed), and  in  recent  years  has  advocated  his  vision  of  a  Semantic  Web. Robert  Cailliau, also  at  CERN,  was  an  early  evangelist  for  the  project.

How  The  Web  Works:

Viewing a  web  page  on  the  World  Wide  Web  normally  begins  either  by  typing  the  URL  of  the  page  into  a  web  browser  or  by  following  a  hyperlink  to  that  page  or  resource.  The  web  browser  then  initiates  a  series  of  communication   messages  behind  the  scenes  in  order  to  fetch  and  display  it.

First  the  server-name  portion  of  the  URL  is  resolved  into  an  IP  address  using  the  global,  distributed  Internet  database  known  as  the  domain  name  system or  DNS. This  IP  address  is  necessary  to  contact  and  send  data  packets  to  the  web  server.

The  browser  then  requests  the  resource  by  sending  an  HTTP  request  to  the  web  server  at  that  particular  address. In  the  case  of  a  typical  web  page  the  HTML  text  of  the  page  is  requested  first  and  parsed  immediately  by  the  web  browser  which  will  then  make  additional  requests  for  images  and  any  other  files  that  form  a  part  of  the  page. Statistics  measuring  a  website’s  popularity  are  usually  based  on  the  number  of ‘ page  views’  or  associated  server  ‘ hits’ or   requests  which  take  place.

Having  received  the  require  files  from  the  web  server  the  browser  renders  the  page  onto  the  screen  as  specified  by  its  HTML,  CSS  and  other  web  languages. Any  images  and  other  resources  are  incorporated  to  produce  the  on-screen  web  page  that  the  user  sees.

Contents  Of  Web  Site:

No  matter  how  great  a  site  looks,  no  amount  of  design  ever  makes  up  for  poor  content. This  is  a  fact  that  many  web  author  lose  sight  of. That’s  why we  have  so  many  sites  around  us  that  offer  the  visitor  the  same  old  thing  – a  bit  of  this,  a  bit  of  that. A  really  good  site  must  have  solid  unique  content. That’s  why  as  exports  recommend  we  started  with  strategy  and  purpose  first – no  with  design. First  off  you  must  quite  clear  of  the  purpose  of  your  site. This  hold  true  for  any  type  of  websites, whether  it’s  a  personal  websites, a small  business  setup, a  hobbyist’s  page  or  e-commerce  or  anything  else. A website  without  purpose  just  takes  space  and  please  no  one  but  its  own  author. So unless  you  are  just  using  your  site  for  storage, start  with  putting  down  your  purpose, your  objectives,  and  the  message.

  Writing  Web  Site:

Writing  for  the  web  is  in  many  ways  different  form  writing  for  print. For  one,  the  reader’s  purpose  in  reading  may  be  different. His  attention  span  is  different. The  reading  experience  online  and  the  way  the  reader’s  eye  moves  across  a  page  are  different. With a printed  page,  there  is  only  one  sort  of  navigation-turn  the  page. But  on  a  web  page, there  can  be  dozens  and  dozens  of  options  all  visible  at  once. And  there’s  your  reader,  finger  poised  over  the  mouse  button,  ready  to  be  interactive. But  with  a  web  page,  interactivity  is  important  because  readers  want  to  do  something. All  this  means  that  information  has  to  be  tailored  and  arranged  specially  for  online  reading. Writing  for  the  web  skillfully  involves  learning  how  to  keep  in  mind  new  online  reading  habits  and  patterns. It  means  being  able  to  put  forward  information  in  a  way  that  draws  the  reader  in  quickly  and  keeps  him  at  the  website  or  at  least  that  it  gives  him  what  he  wants  so  that  he  comes  back  again  and  again.

Web  Site  Styles:

A  static  web  site  is  one  that  has  web  pages  stored  on  the  server  in  the same  form  as  the  user  will  view  them. They  are  edited  using  three  broad  categories  of  software:

Text  editor  such  as  notepad  or  text editor,  where  the  HTML  is  manipulated  directly  within  the  editor  program.

Editor  such  as  Microsoft  Frontpage  and  Macromedia  Dreamweaver  where  the  site  is  edited  using  a  GUI  interface  and  the  underlying  HTML  is  generated  automatically  by  the  editor  software. Template-based  editors  such  as  Rapidweaver  and  iWeb  which  allow  users  to  quickly  create  and  upload    websites  to  a  web  server  without  having  to  know  anything  about  HTML  as  they  just  pick  a  suitable  template  from  a  palette  and  add  pictures  and  text  to  it  in  a  DTP-like  fashion  without  ever  having  to  see  any  HTML  code. A  dynamic  website  is  one  that  has  frequently  changing  information  or  collates  information  on  the  hop  each  time  a  page  is  requested. For  example-  it  would  call  various  bits  of  information  from  a  database  and  put  them  together  in  a  pre-defined  format  to  present  the  reader  with  a  coherent  page. It  interacts  with  users  in  a  variety  of  ways  including  by  reading  cookies  recognizing  user’s  previous  history,  session  variables,  server  side  variables  etc,  or  by  using  direct  interaction  ( form  elements,  mouseovers, etc ). A  site  can  display  the  current  state  of  a  dialog  between  users,  monitor  a  changing  situation  or  provide  information  in  some  way  personalized  to  the  requirements  of  the  individual  users.

What  makes  the  World  Wide  Web  so  exciting  is  the  limitless  ways  in  which  information  and  content  can  be  put  up. Using  color, picture, sounds, movie clips, animation  and  interactivity, you  can  make  sure  your  site  is  compelling  enough  to  draw  visitors  again  and  again  which  is  what  every  website  wants. Naturally, this  makes  the  site’s  design, its  layout,  navigation  and  general  look  and  feel  an  important  aspect  to  work  on. What  usually  tends  to  happen  is  that  people  over-design  a  website,  filling  it  with  bright  starting  colors,  a  feast  of  different  fonts  and  too  many  pictures  for  its  own  good. Many  resource  sites  have  design  is  term  that  is  used  rather  loosely  sometimes  it  can  include  usability  issues, navigation, browser  compatibility  and  so  on. If  your  page  is  about  a  regional  specific  topic, then  make  sure  you  include  that.

Preferably  right  in  the title  of  the  page. Put  the  region  in  the  keywords  and  page  description  as  well. Remember  too  that  even  if  your  topic  is  regional,  it  has  value  to  global  viewers. What  if  someone  from  Germany  is  visiting  your  home  town  and  needs  any  information  there? You  also  might  want  to  expand  your  site  to  give  more  generic  information  that  would  appeal  to  a  more  global  audience.

Language  On  the  Web:

Right  now, most  of  the  pages  on  the  web  are  in  English  but  just  because  you’re  writing  your  page  in  English  in  Australia  does  not  mean  that  a  Canadian  would  understand  it  or  find  it  useful. Make  sure  that  you  avoid  slang  on  your  site  as  that  is  the  most  non-translatable  element  of  a  page. When  you  list  a  price, indicate  what  currency  you’re  using. And  when  you  list  sizes  or  measurements, it helps  if  you  list  conversions  or  link  to  a  conversion  web  site.

Static  Versus  Dynamic:  Static  HTML  sites  have  not  changed  much  since their  development   and  the  advent  of  the  web. Essentially  websites  are  presented  using  a  wide  array  of  tags  that  offer  means  for  usually  laying out  a  site. Search  engines  have  become  very  good  at  recognizing  static  websites. In  general  search  engines  can  navigate  through  a  static  website  very  easily  and  thus  locate  information. However,  there  is  one  significant  disadvantage  of  static  sites,  you  may  need  a  separate  page ( file )  for  every  page  on  your  site. For  example, if  you  want  to  make  a  design  change  that  affects  the  entire  site  you  may  need  to  adjust  all  pages. For  small  sites  this  is  not  a  problem  but  for  large  content  or  e-commerce  sites  creating  new  pages  or  updating  existing  pages  can  be  time  consuming  and  expensive. Certainly  there  is  web  development  software  that  makes  this  a  little  easier  but  in  the  end  static  sites  take  time  to  manage. Interaction  with  visitors  is  a  key  feature  of  the  best  sites  on  the  web. After  all  the  most  popular  computer  operating  systems  in  the  world  may  be  the  ones  used  for  game  playing  machines. It  seems  that  people  hate  the  pickiness  and  precision  of  computer  that   allow  them  to  do  the  things  they  do. On  the  other  hand  they  love  the  illusion  of  the  computer  as  another  person  with  dynamism  comes  interaction. Dynamic  websites  means  that  different  actions  by  the  visitor  cause  different  behaviors  i.e. outputs  by  the  site. That  means  pages  are  created  as the  user  views  the  site. In  most  cases  this  requires  the  use  of  a  database  which  contains  the  site’s  information  and  some  kind of  scripting  setup  that  is  programmed  to  retrieve  the  information  from  the  database.

  What  A  Dynamic  web  Site  can  do?:

building  a  database  driven  web  site  is  one  of  the  best  ways  to  insure  that  your  site  will  grow  into  the  future. Here  are  some  of  the  reasons  why – – –

Manage  Your  Own  Content:  A  database-backed  website  brings  unprecedented  flexibility  to  how  information  is  stored  and  displayed  on  the web. That  means  you  can  add  and  manage  stories,  information, schedules  and  photographs  without  having  to  calla web  master. It’s  a  great  way  to  take  control  of  your  site  while  saving  money  on  maintenance.

Keep  Your  Visitors  Coming  Back:   With  fresh  content  that  you  can  update  at  your  site  will  always  be  relevant. So  instead  of  finding  the  same  stories  and  information  on  your  site, returning  visitors  will  find  information  that’s  new  and  current. Its  easy,  inexpensive  and  will  keep  your  visitors  coming  back  time  and  time  again.

 Grow  into  The  Future:   Building  a  dynamic,  database-driven  site  is  strategically  superior  because  changes  to  the  site  are  incredibly  easy  to  make. Want  a  new  look  on  the  site? No  problem,  since  design  ( presentation )  is  separated  from  the  site’s  content. Need  to  change  content,  that’s  only  a  few  keystrokes  away  with  easy-to-use  administrative  interfaces. Want  to  add  new  pages  or  section’s? not  a  problem  when  you  have  built  your site  on  a  foundation  that’s  both  solid  and   flexible.

Manage  Visitors  Securely:   With  a  data  driven  site  you  can  let  visitors  see  only  information  that  you  want  them  to  see. Build  member’s-only  sections, handle   passwords, lockout  unwanted  requests,  handle  subscription  services, allow  your  staff  access  to  areas  where  others  are  not   allowed. A database-backed  site  can  perform  these  secure  functions  with  ease.

Be  Searchable:   letting  visitors  find  the  information  they  need  quickly  and  easily  is  a  snap  with  a  dynamic  site. Whether  you   are  a  publisher  hosting  thousands  of  articles  or  a  merchant  selling  hundreds  of  widgets,  a  dynamic  site  allows  your  visitors  to  find  what  they  need  in  a  heartbeat.

Harness  Your  Site’s  potential:  Unlike  traditional  “ static ”  sites, a dynamic  site  is  far  more  than  useful  than  simple  “ brochure  ware”. With  dynamic  architecture,  your  site  can  be  put  to  an  infinite  variety  of  valuable  uses. For  example: you  can  easily  connect  a   visitor  with  a near  by  distributor,  connect  a  specific  salesperson  to  a customer or  deliver  an  instant  response  customer  service  request. In short, a  dynamic  site  delivers  more  than  a “ static ”  site  ever  could.

Spend  Less  Time  Managing  Your  Site:  A  dynamic  site  can  reduce  or  eliminate  many  of  the  most  time-consuming  functions   facing  your staffs. That’s  because  many  administrative  functions  can  now  be  automated. For  example:  if  a  deadline  has  passed  or  an  inventory  sold-out, the  site  can  automatically  remove  those  items  from  display. It  could  notify  automatically  and  update  product  pages  on   its  own. Now  your  staff  can  spend  less  time  on  the  web  managing  our  site  and  more  time  doing  the  things  they  do  best.

Handle  Complex  Tasks:   while  dynamic  sites  are  superb for  publishing  and  e-commerce,  they  can  also  be  used  far  complex  tasks  such  as  quoting,  estimating  and  presenting  customized  sales  information  anywhere, any  time. Handling  complex  tasks  is  par  for  the  course  with  a  dynamic  site.

Connect  To  Your  customers:   When  visitors  come  to  your  site, do  you  gather  information  that  can  help  you  serve  them better? With  a  built-in  database,  a  dynamic  site  is  a  natural  for  gathering  customer  preferences. Ask  them  if  they  want   to  subscribe  to  news  letters  or  if  they  are  interested  in  new  products. Test market  new  products. Survey  them  for  valuable  feedback. A  dynamic  web  site  can  help  you  connect  to  customers  in  ways  you  were  never  able  to  before.

Customize  your  message:   is  it   possible  to  respond  on  an  individual  basis  to  an  infinite  number   of  site  visitors? If  you  build  a  dynamic  site,  it  is. From  greeting  customers  individually  after  log-in  to  sending  carefully-crafted  customized  emails, a dynamic  site  will  help  you  send  the  message  that  your  customers  are  more  than  just  numbers.

Developing  A  Dynamic  Site:

In  most  cases  this  requires  the  use  of  a  database  which  contains  the  site’s  information   and  some  kind  of  scripting  setup  that  is  programmed  to  retrieve  the  information  from  the  database. Some  popular  scripting  languages  are  ASP,  ASP.Net,  PHP,  Pearl,  ColdFusion,  JavaScript. Some  popular  databases  are  MySql, MS SQL  Server,  Oracle  etc.

ASP:  Active  Server  Pages,  Microsoft’s  technology  to  enables  HTML  pages  to  be  dynamic  and  interactive  by  embedding  scripts  i.e.  either  VBScript  or  Jscript,  Microsoft’s  alternative  of  JavaScript. Since  the  scripts  in  ASP  pages ( suffix asp )   are  processed  by  the  server,  any  browser  can  work  with ASP  pages  regardless  of  its  Support  for  the  scripting  language  used  there  in.

ASP.Net:  Microsoft  ASP.Net  is  a  set  of  technologies  in  the  Microsoft .NET  Framework  for  building  web  applications  and  XML  web  services. ASP.Net  pages  execute  on  the  server  and  generate  markup  such  as  HTML,  WML  or  XML  that  is  sent  to  a  desktop  or  mobile  browser.  ASP.Net  pages  use  a  compiled, event-driven  programming  model  that  improves  performance  and  enables  the  separation  of  application  logic  and  user  interface. ASP.Net  pages  and  ASP.Net  XML  web  services  files  contain  server-side  logic ( as  opposed  to  client-side  logic ) written  in  Microsoft  Visual  Basic.NET,  Microsoft  Visual  C#.NET  or  any  Microsoft.NET  framework-compatible  language.

PHP:   A recursive  acronym  for  “ hypertext  Preprocessor ”  is  an  open  source  server side  scripting  language  designed  for  creating  robust  and  reliable  dynamic  web  pages  for  e-commerce  and  other  mission  critical  web  applications.

Perl:   Practical  Extraction  and  Reporting  Language  A  robust  programming  language  frequently  used  for  creating  CGI  programs  on  web  servers  because  it  is  faster  than  UNIX  shell  script  programs,  it  can  read  and  write  binary  files  and  it  can  process  very  large  files.

Cold  Fusion:   It  is  an  advanced  website  program  that  runs  on  our  servers  here  at  Thelix. Cold  Fusion  works  in  conjunction  with  a  database  of  information  that  it  draws  from. You  can  use  Cold  Fusion  to  create  dynamic  web  pages that  display  a  variety  of  data,  depending  on  what  the  viewer clicks  on.

JavaScript:   a  scripting  language  from  Netscape  that  is  only  marginally  related  to  java. Java  and  JavaScript  are  not  the  same  thing. JavaScript  was  designed  to  resemble  Java  which  in  turn  looks  a  lot  like  C  and  C++. The  difference  is  that  java  was  built  as  a  general-purpose  object  language  while  JavaScript  is  intended  to  provide  a  quicker  and  simpler  language  in  a  web  page  that  is  interpreted  and  executed  by  the  web  client. The  script  writer  controls  from  within  a  web  document,  often  executed  by  mouse  function,  buttons  or  other  actions  from  the  user. JavaScript  can  be  used  to  fully  control  Netscape  and  Microsoft  web  browsers  including  all  the  familiar  browser  attributes.

MySQL: The   MySQL  database  server  is  the  world’s  most  popular  open  source  database. Over  six  million  installations  use  MySQL  to  power  high-volume  websites  and  other  critical  business  systems – including  industry- leaders like the  associated press, Yahoo, NASA, Sabre  Holdings  and  Suzuki. MySQL  is  an  attractive  alternative  to  higher-cost,  more  complex  database  technology. Its  award-winning  speed, scalability  and  reliability  make  it  the  right  choice  for  corporate  IT  departments, web  developers  and  packaged  software  vendors.

Microsoft  SQL :  Microsoft  SQL  server  is  a  relation  database  management  system  produced  by  Microsoft. It  supports  a dialect of  SQL,  the  most  common  database  language. It is  commonly  used  by  governments  and  businesses  for  small  to  medium  sized  databases  and  completes  with  other  SQL  databases  for  this  market  segment.

Oracle  database:   An  Oracle  database,  strictly  speaking,  is  a  collection  of  data,  is  sometimes  imprecisely  used  to  refer  to  the  DBMS  software  itself. This  error  is  made  in  the  title  of  this  article  and  below. The  oracle  managed  by  an  Oracle  database  management  system  or  DBMS. The  term ” Oracle  database management  system”  can  be  referred  to  without  ambiguity  as  Oracle  DBMS  or  the  databases  which  it  manages  being  relational  in  character, as oracle RDBMS. The  very  useful  distinction  between  data  managed  by  Oracle, an  Oracle  database  and  the  Oracle  RDBMS  is  blurred  by  Oracle  Corporation  itself  when  they  refer  now a days  to  the  Oracle  RDBMS,  the  software  they  sell  to  manage  databases  as  the Oracle  Database. The  distinction  between  the  managed  data ( the  database )  and  the  software  which  manages  the  ( DBMS / RDBMS )  relies, in  Oracle’s  marketing  literature. On  the  capitalization  of  the  world  database. The  Oracle  DBMS  is  produced  and  marketed  by  Oracle  Corporation. The Oracle  DBMS  is  extensively  used  by  many  database  applications  on  most  popular  computing  platforms.

Scripts  &  Database:

We  used  PHP  as  scripting  language  and  MySQL  as  backend  Database. In  the  upcoming  chapter  we  will  discuss  briefly  about  PHP  and  MySQL  and  their  beneficiary  side  over  others  similar  tools.

  Publishing  Web  Page:

 Web  page  production  is  available  to  individuals  outside  the  mass  media. In  order  to  publish  a  web  page,  one  does  not  have  to  go  through  a  publisher  or  other  media  institution  and  potential  readers  could  be  found in  all  corners  of  the  globe. Many  different  kinds  of  information  are available  on the  web  and  for  those  who  wish  to  know  other  societies,  cultures  and  peoples, it  has  become  easier. The  increased  opportunity  to  publish materials  is  observable  in  the  countless  personal  and  social  networking  pages, as  well  as  sites  by families, small  shops  etc.  facilitated  by  the  emergence  of  free  web  hosting  services.

Web  Analytics:

Web  analytics is  the  study  of  the  behavior  of  website  visitors. In a  commercial  context,    web  analytics  especially  refers  to   the  use  of  data  collected  from  a  web  site  to  determine  which  aspects  of  the  website  work  towards  the  business  objectives  for  example  which  landing  pages  encourage  people  to  make  a  purchase.

Data  collected  almost  always  includes  web  traffic  reports. It  may  also  include  e-mail  response  rates, direct  mail  campaign  data,  sales  and  lead  information,  user  performance  data  such as  click  heat  mapping  or  other  custom  metrics  as  needed. This  data  is  typically  compared  against  key  performance  indicators  for  performance  and  used  to  improve  a  web  site  or  marketing  campaign’s  audience  response. Many  different  vendors  provide  web  analyt

Rocky Mountain Spotted Fever

Assignment on web hosting and site promotion, assignment on dna, assignment on recent development in textile fabric wet processing, assignment on gmg airlines in bangladesh, philosophy of mind, annual report 2013-2014 of maruti suzuki india limited, annual report 2009 of bangladesh export import company limited, purple dot, a waitlist and pre-order platform for the fashion industry, rises $4m, presentation on clostridium tetani, latest post, cadmium chloride, improving eye tracking for evaluating brain abnormalities, psychedelics stimulate cells in the hippocampus to alleviate anxiety, sodium iodide – an ionic compound, sodium iodate (naio3), adding depth to the connection between brain structure and ideology.

  • Engineering Mathematics
  • Discrete Mathematics
  • Operating System
  • Computer Networks
  • Digital Logic and Design
  • C Programming
  • Data Structures
  • Theory of Computation
  • Compiler Design
  • Computer Org and Architecture

World Wide Web (WWW)

The World Wide Web (WWW), often called the Web, is a system of interconnected webpages and information that you can access using the Internet. It was created to help people share and find information easily, using links that connect different pages together. The Web allows us to browse websites, watch videos, shop online, and connect with others around the world through our computers and phones.

All public websites or web pages that people may access on their local computers and other devices through the internet are collectively known as the World Wide Web or W3. Users can get further information by navigating to links interconnecting these pages and documents. This data may be presented in text, picture, audio, or video formats on the internet.

What is WWW?

WWW stands for World Wide Web and is commonly known as the Web. The WWW was started by CERN in 1989.  WWW is defined as the collection of different websites around the world, containing different information shared via local servers(or computers).

Web pages are linked together using hyperlinks which are HTML-formatted and, also referred to as hypertext, these are the fundamental units of the Internet and are accessed through Hypertext Transfer Protocol(HTTP) . Such digital connections, or links, allow users to easily access desired information by connecting relevant pieces of information. The benefit of hypertext is it allows you to pick a word or phrase from the text and click on other sites that have more information about it.

History of the WWW

It is a project created, by Tim Berner Lee in 1989, for researchers to work together effectively at CERN. It is an organization, named the World Wide Web Consortium (W3C) , which was developed for further development of the web. This organization is directed by Tim Berner’s Lee, aka the father of the web.  CERN, where Tim Berners worked, is a community of more than 1700 researchers from more than 100 countries. These researchers spend a little time on CERN and the rest of the time they work at their colleges and national research facilities in their home country, so there was a requirement for solid communication so that they can exchange data. 

System Architecture

From the user’s point of view, the web consists of a vast, worldwide connection of documents or web pages. Each page may contain links to other pages anywhere in the world. The pages can be retrieved and viewed by using browsers of which internet explorer, Netscape Navigator, Google Chrome, etc are the popular ones. The browser fetches the page requested interprets the text and formatting commands on it, and displays the page, properly formatted, on the screen. 

The basic model of how the web works are shown in the figure below. Here the browser is displaying a web page on the client machine. When the user clicks on a line of text that is linked to a page on the abd.com server, the browser follows the hyperlink by sending a message to the abd.com server asking it for the page. 

System Architecture

Working of WWW

A Web browser is used to access web pages. Web browsers can be defined as programs which display text, data, pictures, animation and video on the Internet. Hyperlinked resources on the World Wide Web can be accessed using software interfaces provided by Web browsers. Initially, Web browsers were used only for surfing the Web but now they have become more universal.

The below diagram indicates how the Web operates just like client-server architecture of the internet. When users request web pages or other information, then the web browser of your system request to the server for the information and then the  web server   provide requested services to web browser back and finally the requested service is utilized by the user who made the request.

WWW

World Wide Web

Web browsers can be used for several tasks including conducting searches, mailing, transferring files, and much more. Some of the commonly used browsers are Internet Explorer, Opera Mini, and Google Chrome. 

Features of WWW

  • WWW is open source.
  • It is a distributed system spread across various websites.
  • It is a Hypertext Information System. 
  • It is Cross-Platform. 
  • Uses Web Browsers to provide a single interface for many services.
  • Dynamic, Interactive and Evolving.   

Components of the Web

There are 3 components of the web:

  • Uniform Resource Locator (URL): URL serves as a system for resources on the web. 
  • Hyper Text Transfer Protocol (HTTP): HTTP specifies communication of browser and server. 
  • Hyper Text Markup Language (HTML): HTML defines the structure, organisation and content of a web page. 

Difference Between WWW and Internet

Web browser evolution and the growth of the world wide web.

In the early 1990s, Tim Berners-Lee and his team created a basic text web browser. It was the release of the more user-friendly Mosaic browser in 1993 that really sparked widespread interest in the World Wide Web (WWW). Mosaic had a clickable interface similar to what people were already familiar with on personal computers, which made it easier for everyone to use the internet.

Mosaic was developed by Marc Andreessen and others in the United States. They later made Netscape Navigator, which became the most popular browser in 1994. Microsoft’s Internet Explorer took over in 1995 and held the top spot for many years. Mozilla Firefox came out in 2004, followed by Google Chrome in 2008, both challenging IE’s dominance. In 2015, Microsoft replaced Internet Explorer with Microsoft Edge.

The World Wide Web (WWW) has revolutionized how information is accessed and shared globally. It provides a vast network of interconnected documents and resources accessible via the Internet . Through web browsers , users can navigate websites, access multimedia content, communicate, and conduct transactions online. The WWW has transformed communication, commerce, education, and entertainment, shaping modern society and facilitating a connected global community. Its continued evolution and accessibility drive innovation and connectivity worldwide.

Frequently Asked Questions on World Wide Web – FAQs

Can we use the web without an internet .

If a website is on your computer or a computer connected to the same local area network , you can browse it without an internet .

What problems does World Wide Web face?

Our private data is no longer under our control. Misinformation on the internet spreads very easily.

What is the world’s first website?

The first website is info.cern.ch.

What is a URL?

URL stands for Uniform Resource Locator, it is an online address that is given to the web page.

Similar Reads

  • Web technologies

Please Login to comment...

Improve your coding skills with practice.

 alt=

What kind of Experience do you want to share?

GCFGlobal Logo

  • Get started with computers
  • Learn Microsoft Office
  • Apply for a job
  • Improve my work skills
  • Design nice-looking docs
  • Getting Started
  • Smartphones & Tablets
  • Typing Tutorial
  • Online Learning
  • Basic Internet Skills
  • Online Safety
  • Social Media
  • Zoom Basics
  • Google Docs
  • Google Sheets
  • Career Planning
  • Resume Writing
  • Cover Letters
  • Job Search and Networking
  • Business Communication
  • Entrepreneurship 101
  • Careers without College
  • Job Hunt for Today
  • 3D Printing
  • Freelancing 101
  • Personal Finance
  • Sharing Economy
  • Decision-Making
  • Graphic Design
  • Photography
  • Image Editing
  • Learning WordPress
  • Language Learning
  • Critical Thinking
  • For Educators
  • Translations
  • Staff Picks
  • English expand_more expand_less

Internet Basics  - What is the Internet?

Internet basics  -.

What is the Internet?

Internet Basics What is the Internet?

GCFLearnFree Logo

Internet Basics: What is the Internet?

Lesson 1: what is the internet, introduction.

The Internet is an increasingly important part of everyday life for people around the world. But if you've never used the Internet before, all of this new information might feel a bit confusing at first.

Throughout this tutorial, we'll try to answer some basic questions you may have about the Internet and how it's used. When you're done, you'll have a good understanding of how the Internet works , how to connect to the Internet , and how to browse the Web .

The Internet is a global network of billions of computers and other electronic devices. With the Internet, it's possible to access almost any information, communicate with anyone else in the world, and do much more.

You can do all of this by connecting a computer to the Internet, which is also called going online . When someone says a computer is online, it's just another way of saying it's connected to the Internet.

Computers wirelessly connecting to the Internet

What is the Web?

The World Wide Web —usually called the Web for short—is a collection of different websites you can access through the Internet. A website is made up of related text, images, and other resources. Websites can resemble other forms of media—like newspaper articles or television programs—or they can be interactive in a way that's unique to computers.

The purpose of a website can be almost anything: a news platform, an advertisement, an online library, a forum for sharing images, or an educational site like us!

A learner browsing GCFLearnFree.org

Once you are connected to the Internet, you can access and view websites using a type of application called a web browser . Just keep in mind that the web browser itself is not the Internet; it only displays websites that are stored on the Internet.

How does the Internet work?

At this point you may be wondering, how does the Internet work? The exact answer is pretty complicated and would take a while to explain. Instead, let's look at some of the most important things you should know.

It's important to realize that the Internet is a global network of physical cables , which can include copper telephone wires, TV cables, and fiber optic cables. Even wireless connections like Wi-Fi and 3G/4G rely on these physical cables to access the Internet.

When you visit a website, your computer sends a request over these wires to a server . A server is where websites are stored, and it works a lot like your computer's hard drive. Once the request arrives, the server retrieves the website and sends the correct data back to your computer. What's amazing is that this all happens in just a few seconds!

Watch the video below from Tata Communications to learn more about how the Internet functions.

Other things you can do on the Internet

One of the best features of the Internet is the ability to communicate almost instantly with anyone in the world. Email is one of the oldest and most universal ways to communicate and share information on the Internet, and billions of people use it. Social media allows people to connect in a variety of ways and build communities online.

People communicating on the Internet.

There are many other things you can do on the Internet. There are thousands of ways to keep up with news or shop for anything online. You can pay your bills, manage your bank accounts , meet new people, watch TV , or learn new skills. You can learn or do almost anything online.

previous

/en/internetbasics/what-can-you-do-online/content/

  • SEO Tutorial
  • Types of SEO
  • White Hat SEO
  • Black Hat SEO
  • How Search Engine Works
  • What is Domain
  • What is WWW
  • Search Engine vs Portal
  • Google Algorithm Updates

Reseach and Analysis

  • SEO Keyword Research
  • Competitors' Analysis
  • SWOT Analysis
  • Choosing Best Keywords
  • Keyword Research Tools

SEO Guidelines

  • Website Design Guidelines
  • Content Optimization
  • SEO Design and Layout

On Page SEO

  • On Page Optimization
  • SEO Page Title
  • SEO Meta Description
  • SEO Meta Keywords
  • SEO Headings
  • SEO Optimized Domain
  • SEO Canonical Tag
  • SEO Meta Tags
  • Miscellaneous Meta Tags
  • Open Graph Meta Tags
  • SEO Internal link
  • SEO Site Map
  • SEO Hidden Text
  • SEO Web Hosting
  • SEO Robots Meta Tag
  • SEO 301 Redirect
  • SEO 404 error

Off Page SEO

  • Off Page Optimization
  • SEO Page Rank
  • SEO Link Popularity
  • SEO Directory Submission
  • SEO Social Bookmars
  • SEO Blog Submission
  • SEO Article Submission
  • SEO Reciprocial
  • SEO Forum Posting
  • Search Engine Submission
  • RSS Feed Submission
  • Press Release Submission
  • Top 51 Best SEO Tools
  • Off-Page SEO Techniques
  • On-Page SEO Techniques
  • SEMrush Review
  • Semrush Keyword Tool
  • Ahrefs Vs Semrush Vs Moz
  • Semrush Certification
  • Semrush Keyword Research
  • Semrush Training
  • Moz Vs. Semrush
  • Semrush Customer Support
  • Semrush Bot
  • Semrush Toolbar
  • Semrush Keyword Magic Tool
  • Semrush Login
  • Semrush Pricing
  • Semrush Alternatives
  • Semrush Competitors
  • Semrush Free account
  • Semrush Help
  • Spyfu Vs Semrush
  • What is Semrush
  • Pricing Semrush
  • Semrush Affiliate
  • Semrush Domain Authority
  • Semrush Download
  • Semrush Offers
  • Ahrefs vs Semrush
  • Is Semrush Free
  • Semrush Free Trial
  • Semrush Guru
  • Semrush Stock
  • Semrush vs Moz
  • Semrush vs SpyFu
  • SimilarWeb vs Semrush
  • SE Ranking vs Semrush
  • How to use Semrush
  • How to Use Semrush for Keyword Research
  • Neil Patel Vs Semrush
  • Semrush Authority Score
  • Semrush Competitor Analysis
  • Semrush Inc
  • Semrush Plans
  • Semrush Position Tracking
  • Semrush SEO
  • Semrush Site Audit
  • Tools like Semrush
  • Ubersuggest vs Semrush
  • Semrush 30-Day Free Trail
  • Semrush Backlink Audit
  • Semrush Careers
  • Semrush Gratis
  • Semrush Search Volume
  • Semrush SEO Tools
  • Semrush Traffic Analytics
  • SEO PowerSuite vs Semrush
  • What does Semrush do
  • Semrush Account
  • Semrush API Documentation
  • Semrush Backlinks
  • Semrush Discount
  • Semrush Guide
  • Semrush Support
  • Bloggers Passion Semrush
  • Semrush Backlink Checker
  • Semrush Free Alternatives
  • Semrush Guru Trial
  • Semrush PPC Certification
  • What Is Semrush Used For
  • Semrush App
  • Semrush Keywords
  • Semrush Job
  • Semrush Listing Management
  • Semrush Keyword Difficulty
  • Semrush SEO Writing Assistant
  • Semrush Twitter
  • Can I Use Semrush for Free
  • How Much Does Semrush Cost
  • How to Cancel Semrush
  • Semrush Website Traffic
  • Semrush Course
  • Semrush Visibility
  • Semrush IPO

Interview Questions

  • SEO Interview Questions

Latest Courses

We provides tutorials and interview questions of all technology like java tutorial, android, java frameworks

Contact info

G-13, 2nd Floor, Sec-3, Noida, UP, 201301, India

[email protected] .

Latest Post

PRIVACY POLICY

Online Compiler

Want a daily email of lesson plans that span all subjects and age groups?

What is the world wide web.

1,321,359 Views

44,361 Questions Answered

Let’s Begin…

The World Wide Web is used every day by millions of people for everything from checking the weather to sharing cat videos. But what is it exactly? Twila Camp describes this interconnected information system as a virtual city that everyone owns and explains how it’s organized in a way that mimics our brain’s natural way of thinking.

About TED-Ed Animations

TED-Ed Animations feature the words and ideas of educators brought to life by professional animators. Are you an educator or animator interested in creating a TED-Ed Animation? Nominate yourself here »

Meet The Creators

  • Educator Twila Camp
  • Producer Darcy Vorhees
  • Director Tom Beuerlein
  • Script Editor Alex Gendler
  • Narrator Addison Anderson

More from Inventions that Shape History

assignment on world wide web

How did ancient civilizations make ice cream?

Lesson duration 04:55

1,056,629 Views

assignment on world wide web

One of the world’s oldest condiments

Lesson duration 05:20

693,917 Views

assignment on world wide web

How humanity got hooked on coffee

Lesson duration 05:35

1,404,686 Views

assignment on world wide web

What did people do before anesthesia?

Lesson duration 05:27

1,714,809 Views

Session 01 - The Internet, World Wide Web, Web Browsers, Web Sites, and HTML

Harvard Extension School   Fall 2023

Course Web Site: https://cscie12.dce.harvard.edu/

How does the web work?

The internet and the world wide web, how do we make the web work, csci e-12 course, a web address - urls, components of the web, client-side web parts: markup, style, function, html introduction, essential html5 document structure, common html5 elements - a element - anchor, relative urls, "hello world" first assigment, file management, course web hosting server: urls and file locations.

Session 01 - The Internet, World Wide Web, Web Browsers, Web Sites, and HTML, slide1 Welcome!, slide2 How does the web work?, slide3 How does the web work?, slide4 The Internet and the World Wide Web, slide5 The Internet: Schematic, slide6 Tim Berners-Lee on The World Wide Web, slide7 The World Wide Web — key aspects, slide8 How has the web changed? A brief look at Southwest Airlines website evolution, slide9 How do we make the web work?, slide10 How do we make the web work — some tools for our toolbox, slide11 Approaching a Web Project, slide12 CSCI E-12 Course, slide13 CSCI E-12 — Goals for the Course, slide14 CSCI E-12 — Key factors for success in the course, slide15 A Web Address - URLs, slide16 Aside: URLs, URIs, and URNs, slide17 Components of the Web, slide18 Client-side Web Parts: Markup, Style, Function, slide19 Our Solar System: Markup, slide20 Our Solar System: Markup + Style, slide21 Our Solar System: Markup + Style + Function, slide22 HTML Introduction, slide23 Markup - HTML, slide24 Essential HTML5 Document Structure, slide25 HTML Elements - the basic building blocks structure, slide26 HTML Elements - Content can be other HTML elements, slide27 HTML Elements - Sometimes you will have more than one attribute, slide28 HTML Elements that are "empty", slide29 HTML Elements - Sometimes HTML allows you to leave off end tags, slide30 HTML5, slide31 Most commonly used or seen elements, slide32 Learning about HTML elements, slide33 HTML Purpose, slide34 Web Page Structure - header, main, footer, slide35 HTML5 Document Template, slide36 Benefits of Web Standards, slide37 HTML Best Practices to start out with, slide38 Common HTML5 Elements - a element - anchor, slide39 Creating Links, slide40 Relative URLs, slide41 Absolute and Relative Locations, slide42 Relative Paths to Parent Locations, slide43 "Hello World" First Assigment, slide44 File Management, slide45 File Management, slide46 Course Web Hosting Server: URLs and File locations, slide47 Workflow, slide48

Presentation contains 48 slides

CSCI E-12, Fundamentals of Website Development Fall Term 2023 Harvard Extension School

Essential Questions to Consider

Four phases for tonight.

  • A bit about the course
  • Getting started with the web and HTML
  • Orientation to the first assignment

What happens when you enter https://extension.harvard.edu/ into your browser?

harvard extension school home page screenshot

  • Web Browser
  • Web Address (URL) and Communication
  • Web Content (HTML, CSS, JS, images)

web parts

1. Web Browser (HTTP Client)

http client

2. Web Server (HTTP Server)

server-side

3. Communication

Communication between the web browser and web server, including how they communicate (HTTP) and the network.

Visualization of the routing paths of the Internet.

Image credit: Barrett Lyon / The Opte Project "Visualization of the routing paths of the Internet."" Used under the Creative Commons Attribution-NonCommercial 4.0 International License.

The Internet: Schematic

Internet

The Internet came before the Web, and Web "traffic" is not the only type of traffic on the Internet

Tim Berners-Lee on The World Wide Web

Suppose all the information stored on computers everywhere were linked. Suppose I could program my computer to create a space in which everything could be linked to everything. Tim Berners-Lee
The Web evolved into a powerful, ubiquitous tool because it was built on egalitarian principles and because thousands of individuals, universities and companies have worked, both independently and together as part of the World Wide Web Consortium, to expand its capabilities based on those principles. Tim Berners-Lee in Long Live the Web (Scientific American, Nov/Dec 2010)
Today, and throughout this year, we should celebrate the Web’s first 25 years. But though the mood is upbeat, we also know we are not done . We have much to do for the Web to reach its full potential. We must continue to defend its core principles and tackle some key challenges . Tim Berners-Lee in Welcome to the Web's 25 Anniversary (2014)
The web is for everyone, and collectively we hold the power to change it. It won’t be easy. But if we dream a little and work a lot, we can get the web we want. Tim Berners-Lee interview on 30 years of the world wide web in The Guardian (2019)

What is Tim Berners-Lee up to today? He's concerned about personal data sovereignty — and has software and a company to help address it.

The World Wide Web — key aspects

  • HyperText Information System
  • Cross-Platform and Cross-Device ...then and now
  • 255 million unique domains, 201 million active sites, and 12 million "web-facing" computers ( Netcraft Web Server Survey , August 2023)
  • HTML, CSS, JavaScript, HTTP, networking
  • Browsers and engines (Chromium, Webkit, Mozilla), languages (PHP, Python), server software (Apache, nginx, WordPress)
  • Information, Shopping, Banking and Finance, Communication, Business workflows, etc.
  • Dynamic, Interactive, Evolving

How has the web changed? A brief look at Southwest Airlines website evolution

Southwest

  • Physical Desk

Southwest

  • Main categories and functions

Southwest

  • Quick Links

Southwest

  • Travel Tools as icons

Southwest

  • Travel tools as panels
  • Main imagery chages

Southwest

  • Travel Products

Southwest

  • Travel Products expanded, multiple locations
  • Travel tool panels
  • Mega footer beginning

Southwest

  • Less is more
  • "Action" or "Do" panel clearer
  • Social media links

Southwest

Screenshots from my collection and from Internet Archive Wayback Machine

Understand the parts, and how they work individually and how they work together.

  • HTML, CSS, JavaScript
  • Hosting, Web server software, programming languages (JavaScript, Python, etc.)
  • User Experience and Design

This works at multiple layers too —

  • we need to understand HTML, CSS, and JavaScript individually as well as how they interact or relate to one another
  • we also need to understand how we use HTML, CSS, and JavaScript together to accomplish a specific design or interface that will be useful to our users

How do we make the web work — some tools for our toolbox

  • A code editor . Recommendation: Visual Studio Code with a few extensions: Live Server , HTMLHint , Prettier - Code Formatter

I'm not going to ask you to change the browser that you know and use all the time, but I will ask you to have a couple more in the rotation for testing, and to remember that not everyone who views your site will be using your favorite browser.

  • An SFTP client (SFTP = secure file transfer protocol), such as Cyberduck . An SFTP client lets you move files you have edited locally to your web server account.
  • VPN client (VPN = virtual private network). A VPN client lets you connect securely to a network in order to access restricted resources. You will need to be using Harvard VPN in order to SFTP content to your course web hosting account.

There are some more to add to the list eventually, but for now, let's keep it with these!

Approaching a Web Project

elements of user experience

5 Planes from The Elements of User Experience: User-Centered Design for the Web

assignment on world wide web

  • Class & Sections
  • Announcements

CSCI E-12 — Goals for the Course

  • Think like a web developer (programmer, content, project management)
  • Learn syntax of HTML, CSS, JavaScript.
  • Break problems down in parts and build parts up to whole in iterations
  • Learn tools and workflows to improve results and efficiency
  • Understand how things relate to one another.
  • Troubleshoot effectively when things are not working as expected.

CSCI E-12 — Key factors for success in the course

  • Complete and submit the assignments (and to do this you'll likely need to watch the class recordings, attend or view sections, engage in the readings or other resources)
  • Connect with students and course staff
  • Engage with resources available to you: class meetings, sections, Slack, textbook, office hours
  • Start early (even if it is just reading through it)
  • Work to understand a concept outside the context of an assignment — use a 'playground' or 'sandbox' folder on your computer or even something like Codepen or JSFiddle, then apply your understanding to the assignment.
  • leave yourself time to review and revise an assignment
  • Seek out help! Running into roadblocks? Read through thad MDN doc! Check out the text! Attend or watch sections! Communicate early with David or your TA through Slack or Canvas Inbox!

URL = Uniform Resource Locator

"The crucial thing is the URL. The crucial thing is that you can link to anything." — Tim Berners Lee

URL components - scheme, host, and path

URL/URI https://www.archives.gov/historical-docs/voting-rights-act

  • Scheme (also Protocol ) https ://www.archives.gov/historical-docs/voting-rights-act
  • Host (also Authority ) https:// www.archives.gov /historical-docs/voting-rights-act
  • Path https://www.archives.gov /historical-docs/voting-rights-act

Aside: URLs, URIs, and URNs

In the context of the web, all URIs are URLs!

  • URL : Uniform Resource Locator
  • URN : Uniform Resource Name

URI, URN, URL

Huh? A name may be unique, but may not tell you anything about how to locate it.

"Designing Your New Work Life" by Bill Burnett and Dave Evans is a book. It can be uniquely identified by the URN isbn:9780593467459 , and various URLs can be used to locate it (well, if not locate it, at least locate how to purchase or locate it in a library) https://www.amazon.com/Designing-Your-Work-Life-Happiness/dp/0593467450/ https://www.barnesandnoble.com/w/designing-your-new-work-life-bill-burnett/1140537816 https://designingyour.life/designing-your-new-work-life/ https://www.abebooks.com/9780593467459/Designing-New-Work-Life-Thrive-0593467450/plp https://www.worldcat.org/title/1256628247?oclcNum=1256628247

web parts

  • Presentation
  • Manipulations

Our Solar System: Markup

markup

Our Solar System: Markup + Style

markup + style

Our Solar System: Markup + Style + Function

markup + style + function

  • solarsystem.css

Markup - HTML

How a browser displays it.

web page

How Your Browser Thinks About It

dom tree

HTML Elements - the basic building blocks structure

  • Element Name
  • Attribute and Value Pairs

A Hypertext Link

Markup for a Hypertext link:

Harvard Link in a web browser

Start Tag <a href="https://www.harvard.edu/"> Harvard</a>

Element Name < a href="https://www.harvard.edu/">Harvard</a>

Attribute <a href ="https://www.harvard.edu/">Harvard</a>

Attribute Value <a href=" https://www.harvard.edu/ ">Harvard</a>

Content <a href="htts://www.harvard.edu/"> Harvard </a>

End Tag <a href="https://www.harvard.edu/">Harvard </a>

HTML Elements - Content can be other HTML elements

ul and li nodes

ul is an unordered list li is a list item

HTML Elements - Sometimes you will have more than one attribute

img node with two attributes

img is to embed an image

HTML Elements that are "empty"

Note the "end tag" is part of the "start tag" — <link />

img node with two attributes

link is used to reference a CSS stylesheet, a separate document that contains style rules to apply to the HTML document

HTML Elements - Sometimes HTML allows you to leave off end tags

In these cases the "end tags" are "implied" because of what follows.

Learning tip: Always use end tags!

HTML5 Logo

More information: HTML5 Living Standard from the WHATWG . Section 4 contains the List of elements in HTML .

I've highlighted the 23 elements that you will use and/or see most commonly.

  • h1 , h2 , h3 , h4 , h5 , h6

Most commonly used or seen elements

Learning about html elements.

How to find out more about HTML elements?

Two places that I would start are:

HTML Purpose

  • Gives structure and meaning to our content

Think about three aspects of structure :

  • HTML document structure html , head , body
  • Web page structure header , main , nav , footer
  • Content structure Headings ( h1 , h2 , h3 ), lists ( ul and li ), paragraphs ( p ), images ( img ), text, etc.

Example: solarsystem.html

Web Page Structure - header, main, footer

First, recall the basic document structure:

header, main, footer

MDN HTML elements reference: header , main , footer .

HTML5 Document Template

Benefits of web standards.

  • Markup (HTML) Nu Html Checker https://validator.w3.org/nu/
  • Style (CSS) CSS Validation Service https://jigsaw.w3.org/css-validator/
  • Function (JavaScript)
  • Search Engines
  • Forward-compatibility and backward-compatibility.
  • simpler, cleaner pages
  • easier maintenance
  • easier redesign
  • Validation provides baseline when you go to edit.

" Postel's Law " or the " Robustness Principle "

HTML Best Practices to start out with

  • Use start and end tags, even if optional
  • Lower case element and attribute names
  • Use quotes around attribute values

These best practices essentially follow the "XML" syntax rules for HTML5

The anchor — a — element is at the center of the key "hypertext" feature of the web. The a element is how to create hyperlinks from resource to another!

To go along with the a element is the href attribute. The value of the href attribute is the URL that the browser will load when the link is activated (e.g. a mouse click).

The following paragraph was taken from "'Sunshine vitamin' looks a little brighter", Harvard Gazette, February 5, 2013 :

Adequate levels of vitamin D during young adulthood may reduce the risk of adult-onset type 1 diabetes by as much as 50 percent, according to researchers at the Harvard School of Public Health (HSPH). If confirmed in future studies, the findings could lead to a role for vitamin D supplementation in preventing this serious autoimmune disease in adults.

Creating Links

Build confidence by making your links predictable and differentiable .

  • Am I getting 'closer' to my goal?
  • What is the difference between clicking here or clicking there?

scent

  • Link several words or a phrase, not just one or two words
  • Use "title" attribute to elaborate
  • Lie or Mislead
  • " Click Here "
  • Find out more in this knowledge base article

URL https://www.archives.gov/historical-docs/voting-rights-act

  • Scheme https ://www.archives.gov/historical-docs/voting-rights-act
  • Host https:// www.archives.gov /historical-docs/voting-rights-act

Absolute and Relative Locations

  • Where does https://summer.harvard.edu/ go to?
  • How about /images/mug.png ?
  • What about ../styles/site.css ?

Relative locations (URLs) are resolved according to the location (URL) of the containing (starting) document!

Absolute or Fully Qualified URLs

Absolute, or fully-qualified, URLs specify the complete information (scheme, host, port, path).

https://news.harvard.edu/gazette/story/2020/07/public-health-experts-unite-to-bring-clarity-to-coronavirus-response/

Relative or Partial URLs

Relative, or partial, URIs specify partial information. The information not provided is resolved from the current location.

<a href="slide2.html">Slide 2</a>

Relative to Server Root

Is this relative or absolute? Scheme, host, and port would be resolved from current location, but path is absolute

<a href="/copyright.html">copyright information</a>

Relative Paths to Parent Locations

  • ../ refers to the parent directory
  • ./ refers to current directory

Relative links are "transportable":

Hello World - Publishing a web page

  • Get the tools (editor, browser, vpn client, sftp client) in place, and begin to use them.
  • working with ".zip" files with Expand/Compress (Windows) or Unzip/Zip (Mac)
  • keeping everything together in a folder and working with that folder (e.g. "File → Open Folder")
  • Edit an existing HTML document
  • Practice validating your HTML
  • Go through the publishing process on your course web hosting server account
  • Understand how to determine the URL to your published content

For Your Class Work

  • Create a directory or folder for your class work.
  • Create a "playground" or "sandbox" folder where you can play in and experiment in without worrying.
  • Assignments - unzip/extract the materials, then move into your class work folder

For Web Sites

  • Use folders or directories to help organize files. Recommendation is to adopt folder names of styles (for CSS files), scripts (for JavaScript files), and images (for images). . ├── images/ ├── index.html ├── scripts/ └── styles/ └── site.css
  • Use index.html filename as appropriate
  • Prefer filenames that only have lowercase, numeric, underscore or dashes (e.g. avoid spaces, and other things like !@#$%^&*(){}\|?/>

See: Course Web Hosting Server (Dreamhost) on the course web site.

Directory Requests and "index.html"

URL paths that map to a directory. For example the request: https://cwe871.students.cs12.net/big_ideas/ would return the index.html document in the big_ideas directory (e.g. /users/cwe871/public_html/big_ideas/index.html ).

Setup Once for Course

  • Editor . Install Visual Studio Code , with Live Server, W3C Validation, and HTMLHint extensions
  • SFTP client . Install Cyberduck or use Dreamhost File Manager ( https://files.dreamhost.com
  • VPN client . Install Harvard Cisco AnyConnect VPN client software
  • On your computer, create a "course work folder" to keep your work for the course. I recommend something like cscie12-work on your Desktop.

For Assignments

  • Accept the assignment via the GitHub link that creates a repo for you. Then get that repo code locally to edit.
  • Download the assignment ZIP file.
  • Unzip or Extract the ZIP file.
  • Move the folder into your "course work folder" you created above. Move the entire folder.
  • Start Visual Studio Code, and "File → Open Folder" and navigate to the assignment folder to open. In VS Code, I recommend always opening the assignment folder as opposed to opening individual files (treat the assginment folder contents as a unit!)
  • Use VS Code "Live Server" to provide a place to check your work.
  • Edit the HTML, CSS, and/or JavaScript. Save.
  • Periodically check in the browser (a reload may be needed) as well as other checks that may be needed such as validation or accessibility.
  • Repeat the "Edit, Check" cycle until you've satisfied the rubric, yourself, or when the assignment is due.
  • Check in your browser that you can access your work. Copy the URL in your browser, and submit that URL via Canvas
  • Submit the code by giving us the URL to your GitHub repo, or give us the ZIP file of your completed assignment.

The World Wide Web

assignment on world wide web

Tim Berners-Lee, Faculty

Berners-Lee invented the World Wide Web in 1989. His key concepts are familiar to anyone using the Web: HTML (HyperText Markup Language), HTTP (HyperText Transfer Protocol), and URLs (Uniform Resource Locators). When he came to MIT in 1994, he formed the World Wide Web Consortium (W3C) at MIT’s Computer Science and Artificial Intelligence Laboratory to create and maintain open standards for this essential global system. In 2017 he won the Turing Award, the most prestigious honor in computer science.

With your support, we will build a better world.

Share this story

Explore Topics

More Stories

IMAGES

  1. Module 3

    assignment on world wide web

  2. The world wide web

    assignment on world wide web

  3. IT01 Chapter 7

    assignment on world wide web

  4. Tutorial 3

    assignment on world wide web

  5. COMP3322-2022-A1

    assignment on world wide web

  6. The World Wide Web

    assignment on world wide web

VIDEO

  1. Case Study on World Wide Web(1.0.20.30)

  2. World Wide Web In Computer Networks

  3. Definition of world wide Web

  4. UNIT 2 THE WORLD WIDE WEB //Block4 // Media and communication skills #ignou #begg171 #block4unit2

  5. 5.1-Internet And WWW

  6. When was the World Wide Web (WWW) invented?

COMMENTS

  1. The birth of the Web

    Learn how British scientist Tim Berners-Lee invented the World Wide Web in 1989 while working at CERN. Discover the history and impact of the first website and the web software released by CERN.

  2. What's the difference between the Internet and the World Wide Web

    The Internet is a network of networks, while the World Wide Web is the system of using hypertext to access the Internet. Learn the history, the difference and the examples of both terms.

  3. Difference Between Internet and WWW

    Learn the difference between internet and WWW, two terms that are often used interchangeably but have distinct meanings. Internet is a global network of networks, while WWW is a service that provides access to information via hyperlinks.

  4. History of the Internet and World Wide Web (WWW)

    THE WORLD WIDE WEB. Perhaps the invention that most facilitated the growth of the Internet as a global information-sharing system is the World Wide Web. Unlike the Internet, however, the early design and development of the World Wide Web was primarily the doing of just one person: Tim Berners-Lee. Working as a contract programmer at the Geneva ...

  5. Development of the Internet and the World Wide Web

    This chapter traces the development of the Internet and the World Wide Web, focusing on the federal role in funding research and supporting the deployment of networking infrastructure. It covers four periods of development, from 1960 to 1995, and highlights the key projects and innovations that shaped the Internet.

  6. World Wide Web

    Learn how to search for materials on the web, make bookmarks, and download multimedia files for foreign language teaching. Find out the name, stars, and price of the hotel in the southernmost city of the world.

  7. The World Wide Web from its origins

    This web page provides the solution for a reading passage about the history and impact of the World Wide Web, invented by Tim Berners-Lee in 1989. It also explains the questions and answers for the IELTS reading test based on the passage.

  8. Internet vs World Wide Web

    The World Wide Web (WWW) is one set of software services running on the Internet. The Internet itself is a global, interconnected network of computing devices. This network supports a wide variety of interactions and communications between its devices. The World Wide Web is a subset of these interactions and supports websites and URIs.

  9. World Wide Web Quizzes, Questions & Answers

    Test your knowledge about the World Wide Web, the system of linked hypertext documents on the Internet. Find quizzes on the history, technology, applications, and challenges of the Web.

  10. Assignment on World Wide Web

    The World Wide Web was created in 1989 by Sir Tim Berners Lee, working at CERN in Geneva, Switzerland. Since then, Berners Lee has played an active role in guiding the development of web standards ( such as the markup languages in which web pages are composed), and in recent years has advocated his vision of a Semantic Web.

  11. World Wide Web (WWW)

    The World Wide Web (WWW), often called the Web, is a system of interconnected webpages and information that you can access using the Internet. It was created to help people share and find information easily, using links that connect different pages together. The Web allows us to browse websites, watch videos, shop online, and connect with ...

  12. World Wide Web

    The World Wide Web is an information system that enables content sharing over the Internet through user-friendly ways. It was invented by Tim Berners-Lee in 1989 and opened to the public in 1991, using HTTP protocol and HTML language.

  13. World Wide Web

    The World Wide Web is the leading information retrieval service of the Internet, based on hyperlinks that connect documents and multimedia. Learn about its development, features, uses, and challenges from Britannica's editors and articles.

  14. World Wide Web (WWW)

    What will Web 3.0 look like compared to Web 1.0 and 2.0? The World Wide Web continues to evolve. The first generation of the Web, Web 1.0, which Berners-Lee originally defined in 1989, had no video content and a page format similar to that of a printed page. Web 1.0 was primarily static and focused on providing information.

  15. Internet Basics: What is the Internet?

    Learn the definition, history, and features of the Internet, a global network of billions of devices. Find out how to access and use the Web, email, social media, and more on the Internet.

  16. What is World Wide Web?

    World Wide Web is a collection of websites or web pages connected by hyperlinks and accessed by HTTP. Learn how it was invented by Tim Berners-Lee in 1989 and how it differs from internet.

  17. What is the World Wide Web?

    The World Wide Web is used every day by millions of people for everything from checking the weather to sharing cat videos. But what is it exactly? Twila Camp describes this interconnected information system as a virtual city that everyone owns and explains how it's organized in a way that mimics our brain's natural way of thinking.

  18. Session 01

    Communication between the web browser and web server, including how they communicate (HTTP) and the network. The Internet and the World Wide Web. Image credit: Barrett Lyon / The Opte Project "Visualization of the routing paths of the Internet."" Used under the Creative Commons Attribution-NonCommercial 4.0 International License. The Internet ...

  19. The World Wide Web

    The World Wide Web. Tim Berners-Lee, Faculty. Berners-Lee invented the World Wide Web in 1989. His key concepts are familiar to anyone using the Web: HTML (HyperText Markup Language), HTTP (HyperText Transfer Protocol), and URLs (Uniform Resource Locators). When he came to MIT in 1994, he formed the World Wide Web Consortium (W3C) at MIT's ...

  20. Lesson: The World Wide Web

    The World Wide Web is just one of the services provided by the internet. You don't need an internet connection for the World Wide Web. Q2. Who created the internet? Bill Gates. Google. The British government. Correct answer: The United States military. The United States military. Q3.