INTRODUCTION TO WEB TECHNOLOGY (TIT-503) UNIT I
Introduction and Web Development Development Strategies History of Web, Protocols governing Web, Creating Websites for individual and Corporate World, Cyber Laws Web Applications, Writing Web Projects, Identification of Objects, Target Users, Web Team, Planning and Process Development.
INTRODUCTION AND WEB DEVELOPMENT STRATEGIES / WEB TEAM
Internet or commonly known as WEB is defined as a network of networks. The statement ‘NETWORK OF NETWORK’ contains a hidden definition in itself. As we know that in the early stage of development in networks only homogenous systems were able to communicate. But, as the technology has grown, new technology device devicess and softwa software re had emerge emerged d which which allow allow hetero heterogen geneou eouss network to behave like a common group. Internet is collection of such heterogeneous/homogeneous networks. The technologies in int interne ernett allow low one one net network work to comm ommuni unicat cate with ith anot nother her transparently. These days internet is covering almost all aspects of humans daily life and therefore well defined strategies are required to develop as well as use this emerging technology. Emerging of
E-commerce and it’s vast use by banks and other corporate had lead lead to thin think k abou aboutt thes thesee deve develo lopm pmen entt stra strate tegi gies es a lot. lot. Thes Thesee development and use is under a law commonly known as CYBER LAW
(Will
be
dealing
in
detail
later
on)
and
orga organi niza zati tion ons/ s/in indi divi vidu dual alss are are boun bound d to foll follow ow thes thesee rule ruless and and regulations. Prior to the widespread inter-networking that led to the Internet, most communication networks were limited by their nature to only allow communications between the stations on the network, and the the prev preval alen entt comp comput uter er netw networ orki king ng meth method od was was base based d on the the centra centrall mainfr mainframe ame method method.. In the 1960s, 1960s, comput computer er resear researche chers, rs, Levi C. Finch and Robert W. Taylor pioneered calls for a joinedup
global
network
to
address
interoperabili ility
problems.
Conc Concur urre rent ntly ly,, seve severa rall rese resear arch ch prog progra rams ms bega began n to rese resear arch ch principles of networking between separate physical networks, and this led to the development of Packet switching. These included Dona Donald ld Davi Davies es (NPL (NPL), ), Paul Paul Bara Baran n (RAN (RAND D Corp Corpor orat atio ion) n),, and and Leonard Kleinrock's MIT and UCLA research programs.
This led to the development of several packet switched networking solutions in the late 1960s and 1970s, including ARPANET, and X.25. Additionally, public access and hobbyist networking systems grew grew in popula popularit rity, y, includ including ing UUCP. UUCP. They They were were howeve howeverr still still disjoi disjointe nted d separa separate te networ networks, ks, served served only only by limite limited d gatewa gateways ys between networks. This led to the application of packet switching to deve develo lop p a prot protoc ocol ol for for inte interr-ne netw twor orki king ng,, wher wheree mult multip iple le different networks could be joined together into a super-framework of networks. By defining a simple common network system, the Inte Intern rnet et prot protoc ocol ol suit suite, e, the the conc concep eptt of the the netw networ ork k coul could d be separated from its physical implementation. This spread of internetwork network began to form into the idea of a global global inter-netw inter-network ork that would be called 'The Internet', and this began to quickly spread as existing networks were converted to become compatible with this. This This spre spread ad quic quickl kly y acro across ss the the adva advanc nced ed tele teleco comm mmuni unica cati tion on networks of the western world, and then began to penetrate into the rest of the world as it became the de-facto international standard
and global network. However, the disparity of growth led to a Digital divide that is still a concern today. Follow Following ing commer commercia cializ lizati ation on and introd introduct uction ion of privat privately ely run Internet Service Providers in the 1980s, and its expansion into popular use in the 1990s, the Internet has had a drastic impact on cult cultur uree and and comm commer erce ce.. This This incl includ udes es the the rise rise of near near inst instan antt communication by e-mail, text based discussion forums, the World Wide Web. Investor speculation in new markets provided by these innovations would also lead to the inflation and collapse of the Dot-com bubble, a major market collapse. But despite this, growth of the Internet continued, and still does.
HISTORY
OF
WEB
AND
WEB
GOVERNING
PROTOCOLS
In the the 1950 1950ss and and earl early y 1960 1960s, s, prio priorr to the the wide widesp spre read ad inte interrnetworking that led to the Internet, most communication networks were were limi limite ted d by thei theirr natu nature re to only only allo allow w comm commun unic icat atio ions ns between the stations on the network. Some networks had gateways
or bridges between them, but these bridges were often limited or bui built lt spec specif ific ical ally ly for for a sing single le use. use. One One prev preval alen entt comp comput uter er networking method was based on the central mainframe method, simply allowing its terminals to be connected via long leased lines. This method was used in the 1950s by Project RAND to support researchers such as Herbert Simon, in Pittsburgh, Pennsylvania, when when coll collab abor orat atin ing g acro across ss the the cont contin inen entt with with rese resear arch cher erss in Sulliv Sullivan, an, Illino Illinois, is, on automa automated ted theore theorem m provin proving g and artif artifici icial al intelligence. In Octo Octobe berr 1962 1962,, Lick Lickli lide derr was was appo appoin inte ted d head head of the the Unit United ed Stat States es Depa Depart rtme ment nt of Defe Defens nse' e'ss Adva Advanc nced ed Rese Resear arch ch Proj Projec ects ts Agency, now known as DARPA, within the information pro proce cess ssin ing g offi office ce.. Ther Theree he form formed ed an info inform rmal al grou group p with within in DARPA to further computer research. As part of the information processing office's role, three network terminals had been installed: one for System Development Corporation in Santa Monica, one for Project Genie at the University of California, Berkeley and one for the Compatible Time-Sharing System project at the Massachusetts
Institute of Technology (MIT). Licklider's identified need for internetworking would be made obviously evident by the problems this caused. At the the tip tip of the the inte interr-ne netw twor orki king ng prob proble lem m lay lay the the issu issuee of conn connec ecti ting ng sepa separa rate te phys physic ical al netw networ orks ks to form form one one logi logica call network, with much wasted capacity inside the assorted separate networks. During the 1960s, Donald Davies (NPL), Paul Baran (RAND (RAND Corpor Corporat ation ion), ), and Leonar Leonard d Kleinr Kleinrock ock (MIT) (MIT) develo developed ped and implemented packet switching. The notion that the Internet was developed to survive a nuclear attack has its roots in the early the theorie oriess deve devellope oped by RAND AND, but but is an urba urban n legend gend,, not not supp suppor orte ted d by any any Inte Intern rnet et Engi Engine neer erin ing g Task Task Forc Forcee or othe other r document. Early networks used for the command and control of nucl nuclea earr forc forces es were were mess messag agee swit switch ched ed,, not not pack packet et-s -swi witc tche hed, d, although current strategic military networks are, indeed, packetswitch switching ing and connec connecti tionl onless ess.. Baran' Baran'ss resea research rch had approa approache ched d packet switching from studies of decentralisation to avoid combat damage compromising the entire network.
Prom Promot oted ed to the the head head of the the info inform rmat atio ion n proc proces essi sing ng offi office ce at DARPA, Robert Taylor intended to realize Licklider's ideas of an interconnected networking system. Bringing in Larry Roberts from MIT, he initiated a project to build such a network. The first ARPA ARPANE NET T link link was was esta establ blis ishe hed d betw betwee een n the the Univ Univer ersi sity ty of California, Los Angeles and the Stanford Research Institute on 22:30 hours on October 29, 1969. By 5 December 1969, a 4-node network was connected by adding the University of Utah and the Univ Univer ersi sity ty of Cali Califo forn rnia ia,, Sant Santaa Barb Barbar ara. a. Buil Buildi ding ng on idea ideass developed in ALOHAnet, the ARPANET grew rapidly. By 1981, the number of hosts had grown to 213, with a new host being added approximately every twenty days. ARPANET became the technical core of what would become the Internet, and a primary tool in developing the technologies used. ARPANE ARPANET T develo developme pment nt was was cente centered red around around the Reques Requestt for Comm Commen ents ts (RFC (RFC)) proc proces ess, s, stil stilll used used toda today y for for prop propos osin ing g and and distributing Internet Protocols and Systems. RFC 1, entitled "Host Software", was written by Steve Crocker from the University of
California, Los Angeles, and published on April 7, 1969. These early years were documented in the 1972 film Computer Networks: The Heralds of Resource Sharing. Inte Intern rnat atio iona nall coll collab abor orat atio ions ns on ARPA ARPANE NET T were were spar sparse se.. For For variou variouss politi politica call reason reasons, s, Europe European an develo developer perss were were concer concerned ned with developing the X.25 networks. Notable exceptions were the Norwegian Seismic Array (NORSAR) in 1972, followed in 1973 by Sweden with satellite links to the Tanum Earth Station and University College London.
X.25 AND PUBLIC ACCESS
Main articles: X.25, Bulletin board system, and FidoNet Following on from ARPA's research, packet switching network standards were developed by the International Telecommunication Union (ITU) in the form of X.25 and related standards. In 1974, X.25 formed the basis for the SERCnet network between British academ academic ic and resear research ch sites, sites, which which later later became became JANET. JANET. The
initial ITU Standard on X.25 was approved in March 1976. This standard was based on the concept of virtual circuits.
The British Post Office, Western Union International and Tymnet coll collab abor orat ated ed to crea create te the the firs firstt inte intern rnat atio iona nall pack packet et swit switch ched ed network, referred to as the International Packet Switched Service (IPSS), in 1978. This network grew from Europe and the US to cover Canada, Hong Kong and Australia by 1981. By the 1990s it provided a worldwide networking infrastructure.[7]
Unlike ARPAnet, X.25 was also commonly available for business use. Telenet offered its Telemail electronic mail service, but this was oriented to enterprise use rather than the general email of ARPANET. The first dial-in public networks used asynchronous TTY terminal protocols to reach a concentrator operated by the public network. Som Some publ public ic netw networ orks ks,, such such as Com CompuS puServe erve use used X.25 .25 to mult multip iple lex x the the term termin inal al sess sessio ions ns into into thei theirr
pack packet et-s -swi witc tche hed d
bac backb kbon ones es,, whil whilee othe others rs,, such such as Tymn Tymnet et,, used used prop propri riet etar ary y protocols. In 1979, CompuServe became the first service to offer elec electr tron onic ic mail mail capa capabi bili liti ties es and and tech techni nica call supp suppor ortt to pers person onal al computer computer users. The company company broke new ground again in 1980 as the first to offer real-time chat with its CB Simulator. There were also the America Online (AOL) and Prodigy dial in networks and many many bullet bulletin in board board syste system m (BBS) (BBS) networ networks ks such such as FidoNe FidoNet. t. FidoNe FidoNett in partic particula ularr was was popula popularr among amongst st hobbyi hobbyist st comput computer er users, many of them hackers and amateur radio operators.
UUCP( Main articles: UUCP and Usenet ) In 1979, two students at Duke University, Tom Truscott and Jim Ellis, came up with the idea of using simple Bourne shell scripts to transfer news and messages on a serial line with nearby University of North Carolina at Chapel Hill. Following public release of the software, the mesh of UUCP hosts forwarding on the Usenet news rapi rapidl dly y expa expand nded ed.. UUCP UUCPne net, t, as it woul would d late laterr be name named, d, also also create created d gatewa gateways ys and links links betwee between n FidoNe FidoNett and dial-u dial-up p BBS
host hosts. s. UUCP UUCP netw networ orks ks spre spread ad quic quickl kly y due due to the the lowe lowerr cost costss involved, and ability to use existing leased lines, X.25 links or even ARPANET connections. By 1981 the number of UUCP hosts had grown to 550, nearly doubling to 940 in 1984.
Merging the networks and creating the Internet (TCP/IP)
INTERNET PROTOCOL SUITE With so many different network methods, something was needed to unif unify y the them. Rob Robert ert E. Kahn of DARP ARPA and ARPA RPANET recruited Vinton Cerf of Stanford University to work with him on the problem. By 1973, they had soon worked out a fundamental reformulation, where the differences between network protocols were hidden by using a common internetwork protocol, and instead of the netw networ ork k bei being res respons ponsiible ble for rel reliabi iabili litty, as in the the ARPANE ARPANET, T, the hosts hosts becam becamee respon responsib sible. le. Cerf Cerf credit creditss Huber Hubertt Zimmerman, Gerard LeLann and Louis Pouzin (designer of the CYCLADES network) with important work on this design.[8]
At this time, the earliest known use of the term Internet was by Vinton Cerf, who wrote: “Specification of Internet Transmission Control Program”. With the role of the network reduced to the bare minimum, it became possible to join almost any networks together, no matter what their char charac acte teri rist stic icss were were,, ther thereb eby y solv solvin ing g Kahn Kahn's 's init initia iall prob proble lem. m. DARPA agreed to fund development of prototype software, and after several years of work, the first somewhat crude demonstration of a gateway between the Packet Radio network in the SF Bay area and the ARPANET was conducted. On November 22, 1977 a three network demonstration was conducted including the ARPANET, the the Pack Packet et Radi Radio o Netw Networ ork k and and the the Atla Atlant ntic ic Pack Packet et Sate Satell llit itee netw networ ork— k—al alll spon sponso sore red d by DARP DARPA. A. Stem Stemmi ming ng from from the the firs firstt specifications of TCP in 1974, TCP/IP emerged in mid-late 1978 in near nearly ly fina finall form form.. By 1981 1981,, the the asso associ ciat ated ed stan standa dard rdss were were published as RFCs 791, 792 and 793 and adopted for use. DARPA sponsored
or
encouraged
the
development
of
TCP/IP
implementations for many operating systems and then scheduled a migration of all hosts on all of its packet networks to TCP/IP. On 1 Janu Januar ary y 1983 1983,, TCP/ TCP/IP IP prot protoc ocol olss beca became me the the only only appr approv oved ed protocol on the ARPANET, replacing the earlier NCP protocol. ARPANET to Several Federal Wide Area Networks: MILNET, NSI, and NSFNet
ARPANET and NSFNet After the ARPANET had been up and running for several years, ARPA looked for another agency to hand off the network to; ARPA's primary mission was funding cutting edge research and development, not running a communications utility. Eventually, in July July 1975 1975,, the the netw networ ork k had had been been turn turned ed over over to the the Defe Defens nsee Communications Agency, also part of the Department of Defense. In 1983, the U.S. military portion of the ARPANET was broken off as a separate network, the MILNET. MILNET subsequently became the unclassified but military-only NIPRNET, in parallel with the SECRET-level SIPRNET and JWICS for TOP SECRET
and above. NIPRNET does have controlled security gateways to the public Internet. The netw networ orks ks base based d arou around nd the the ARPA ARPANE NET T were were gove govern rnme ment nt funded funded and theref therefore ore restri restricte cted d to noncom noncomme merci rcial al uses uses such such as research; unrelated commercial use was strictly forbidden. This initially restricted connections to military sites and universities. During the 1980s, the connections expanded to more educational institutions, and even to a growing number of companies such as Digital Equipment Corporation and Hewlett-Packard, which were participating in research projects or providing services to those who were. Seve Severa rall othe otherr bran branch ches es of the the U.S. U.S. gove govern rnme ment nt,, the the Nati Nation onal al Aerona Aeronauti utics cs and Space Space Agency Agency (NASA (NASA), ), the Nation National al Scienc Sciencee Foundation (NSF), and the Department of Energy (DOE) became heavily involved in internet research and started development of a succ succes esso sorr to ARPA ARPANE NET. T. In the the mid mid 1980 1980ss all all thre threee of thes thesee bra branc nche hess deve develo lope ped d the the firs firstt Wide Wide Area Area Netw Networ orks ks base based d on TCP/I TCP/IP. P. NASA NASA deve develo lope ped d the the NASA NASA Scie Scienc ncee Netw Networ ork, k, NSF NSF
deve develo lope ped d CSN CSNET and DOE evol evolve ved d the Ene Energy rgy Scie Scienc nces es Network or ESNet. More More explic explicitl itly, y, NASA NASA develo developed ped a TCP/IP TCP/IP based based Wide Wide Area Area Net Netwo work rk,, NASA NASA Scie Scienc ncee Netw Networ ork k (NSN (NSN), ), in the the mid mid 1980 1980ss conn connec ecti ting ng spac spacee scie scient ntis ists ts to data data and and info inform rmat atio ion n stor stored ed anywhere in the world. In 1989, the DECnet-based Space Physics Analysis Network (SPAN) and the TCP/IP-based NASA Science Network (NSN) were brought together at NASA Ames Research Center creating the first multiprotocol wide area network called the NASA Science Internet, or NSI. NSI was established to provide a tota totall inte integr grat ated ed comm commun unic icat atio ions ns infr infras astru truct ctur uree to the the NASA NASA scientifi scientificc community community for the advancement advancement of earth, earth, space and life sciences. sciences. As a high-speed high-speed,, multiproto multiprotocol, col, internati international onal network, network, NSI NSI provid provided ed connec connectiv tivity ity to over over 20,000 20,000 scient scientist istss across across all seven continents. In 1984 NSF developed CSNET exclusively based on TCP/IP. CSNET connected with ARPANET using TCP/IP, and ran TCP/IP over X.25, but it also supported departments without sophisticated
network connections, using automated dial-up mail exchange. This grew into the NSFNet backbone, established in 1986, and intended to connect and provide access to a number of supercomputing centers established by the NSF.[12] TRANSITION TOWARD AN INTERNET
The term "Internet" was adopted in the first RFC published on the TCP protocol (RFC 675: Internet Transmission Control Program, December 1974). It was around the time when ARPANET was interlinked with NSFNet, that the term Internet came into more general use,[14] with "an internet" meaning any network using TCP/IP. "The Internet" came to mean a global and large network using TCP/IP. Previously "internet" and "internetwork" had been used interchangeably, and "internet protocol" had been used to refer to other networking systems such as Xerox Network Services. As interest in wide spread networking grew and new applications for it arrived, the Internet's technologies spread throughout the rest of the world. TCP/IP's network-agnostic approach meant that it was easy to use any existing network infrastructure, such as the
IPSS X.25 network, to carry Internet traffic. In 1984, University Coll Colleg egee Lond London on repl replac aced ed its its tran transa satl tlan anti ticc sate satell llit itee link linkss with with TCP/IP over IPSS. Many sites unable to link directly to the Internet started to create simple gateways to allow transfer of e-mail, at that time the most impo import rtaant
appl pplicat icatiion. on.
Site Sitess
whi which
only only
had had
inte ntermit rmitttent ent
connections used UUCP or FidoNet and relied on the gateways between these networks and the Internet. Some gateway services went beyond simple e-mail peering, such as allowing access to FTP sites via UUCP or e-mail. TCP/IP BECOMES WORLDWIDE
The first ARPANET connection outside the US was established to NORSAR in Norway in 1973, just ahead of the connection to Great Britain. These links were all converted to TCP/IP in 1982, at the same time as the rest of the Arpanet.
[edit] CERN, the European internet, the link to the Pacific and beyond
Between 1984 and 1988 CERN began installation and operation of TCP/I TCP/IP P to inte interc rcon onne nect ct its its majo majorr inte intern rnal al comp comput uter er syst system ems, s, work workst stat atio ions ns,, PC's PC's and and an acce accele lera rato torr cont contro roll syst system em.. CERN CERN contin continued ued to operat operatee a limite limited d self-d self-deve evelop loped ed syste system m CERNET CERNET internally and several incompatible (typically proprietary) network protocols externally. There was considerable resistance in Europe towards more widespread use of TCP/IP and the CERN TCP/IP intranets remained isolated from the Internet until 1989. In 1988 Daniel Karrenberg, from CWI in Amsterdam, visited Ben Segal, CERN's TCP/IP Coordinator, looking for advice about the tran transi siti tion on of the the Euro Europe pean an side side of the the UUCP UUCP Usen Usenet et netw networ ork k (much of which ran over X.25 links) over to TCP/IP. In 1987, Ben Segal had met with Len Bosack from the then still small company Cisco about purchasing some TCP/IP routers for CERN, and was able to give Karrenberg advice and forward him on to Cisco for the appropriate hardware. This expanded the European portion of the Internet across the existing UUCP networks, and in 1989 CERN
opened its first external TCP/IP connections. This coincided with the creation of Réseaux IP Européens (RIPE), initially a group of IP netw networ ork k admi admini nist stra rato tors rs who who met met regu regula larl rly y to carr carry y out out cocoordi ordina nati tion on work work toge togeth ther er.. Late Later, r, in 1992 1992,, RIPE RIPE was was form formal ally ly registered as a cooperative in Amsterdam. At the same time as the rise of internetworking in Europe, ad hoc netw networ orki king ng to ARPA ARPA and and in-be in-betw twee een n Aust Austra rali lian an univ univer ersi siti ties es formed, based on various technologies such as X.25 and UUCPNet. These were limited in their connection to the global networks, due to the cost of making individual international UUCP dial dial-u -up p or X.25 X.25 conn connec ecti tion ons. s. In 1989 1989,, Aust Austra rali lian an univ univer ersi siti ties es joi joine ned d the pus push tow towards rds usin using g IP prot protoc ocol olss to uni unify their heir networking infrastructures. AARNet was formed in 1989 by the Australian Vice-Chancellors' Committee and provided a dedicated IP based network for Australia. The Internet began to penetrate Asia in the late 1980s. Japan, whic which h had had buil builtt the the UUCP UUCP-b -bas ased ed netw networ ork k JUNE JUNET T in 1984 1984,, connected to NSFNet in 1989. It hosted the annual meeting of the
Inte Intern rnet et Socie ociety ty,,
INET INET''92, 92,
in
Kobe Kobe..
Singa ingapo pore re deve devellope oped
TEC TECHNET HNET in 1990 1990,, and and Thail hailan and d gai gained ned a glob globaal Inte Intern rneet connec connectio tion n betwe between en Chulal Chulalong ongkor korn n Univer Universit sity y and UUNET UUNET in 1992. DIGITAL DIVIDE While developed countries with technological infrastructures were joining the Internet, developing countries began to experience a digital divide separating them from the Internet. On an essentially cont contin inen enta tall basi basis, s, they they are are buil buildi ding ng orga organi niza zati tion onss for for Inte Intern rnet et resour resource ce admini administr strati ation on and sharin sharing g operat operation ional al experi experienc ence, e, as more and more transmission facilities go into place. AFRICA
At the beginning of the 1990s, African countries relied upon X.25 IPSS and 2400 baud modem UUCP links for international and internetwork computer communications. In 1996 a USAID funded pro proje ject ct,, the the Lela Leland nd init initia iati tive ve,, star starte ted d work work on deve develo lopi ping ng full full Intern Internet et connec connectiv tivity ity for the contin continent ent.. Guinea Guinea,, Mozamb Mozambiqu ique, e,
Madagascar and Rwanda gained satellite earth stations in 1997, followed by Côte d'Ivoire and Benin in 1998. Africa
is
building
an
Internet
infrastructure.
AfriNIC,
headquartered in Mauritius, manages IP address allocation for the continent. As do the other Internet regions, there is an operational foru forum, m, the the Inte Intern rnet et Comm Commun unit ity y of Opera perati tion onal al Netw Networ orki king ng Specialists. The There are a wide wide rang rangee of prog progra rams ms bot both to prov provid idee high high- perfo performa rmance nce transm transmiss ission ion plant, plant, and the wester western n and southe southern rn coasts have undersea optical cable. High-speed cables join North Africa and the Horn of Africa to intercontinental cable systems. Undersea cable development is slower for East Africa; the original joint joint effort effort betwee between n New Partne Partnersh rship ip for Africa Africa's 's Devel Developm opment ent (NEP (NEPAD AD)) and and the the East East Afri Africa ca Subm Submar arin inee Syst System em (Eas (Eassy sy)) has has broken off and may become two efforts. ASIA AND OCEANIA
The The
Asia
Pacific
Network
Information
Centre
(APNIC),
headquartered in Australia, manages IP address allocation for the
continent. APNIC sponsors an operational forum, the Asia-Pacific Regi Region onal al Inte Intern rnet et Conf Confer eren ence ce on Oper Operat atio iona nall
Tech Techno nolo logi gies es
(APRICOT). In 1991 1991,, the the Peop People le's 's Repu Republ blic ic of Chin Chinaa saw saw its its firs firstt TCP/ TCP/IP IP college network, Tsinghua University's TUNET. The PRC went on to make its first global Internet connection in 1995, between the Bei Beijing jing
Elec Electr troo-Sp Speectro ctrom meter eter
Col Collabo labora rati tion on
and
Stan Stanfford
University's Linear Accelerator Center. However, China went on to implement its own digital divide by implementing a countrywide content filter. LATIN AMERICA
As with with the the othe otherr regi region ons, s, the the Lati Latin n Amer Americ ican an and and Cari Caribb bbea ean n Internet Addresses Registry (LACNIC) manages the IP address space and other resources for its area. LACNIC, headquartered in Uruguay, operates DNS root, reverse DNS, and other key services. OPENING THE NETWORK TO COMMERCE
The interest in commercial use of the Internet became a hotly debated topic. Although commercial use was forbidden, the exact
defini definitio tion n of commer commercia ciall use could could be unclea unclearr and subjec subjectiv tive. e. UUCP UUCPNe Nett and and the the X.25 X.25 IPSS IPSS had had no such such rest restri rict ctio ions ns,, whic which h woul would d even eventu tual ally ly see see the the offi offici cial al barr barrin ing g of UUCP UUCPNe Nett use use of ARPA ARPANE NET T and and NSFN NSFNet et conn connec ecti tion ons. s. Some Some UUCP UUCP link linkss stil stilll remained connecting to these networks however, as administrators cast a blind eye to their operation. During During the late late 1980s, 1980s, the first first Intern Internet et servic servicee provid provider er (ISP) (ISP) comp compan anie iess were were form formed ed.. Comp Compan anie iess like like PSIN PSINet et,, UUNE UUNET, T, Netcom, and Portal Software were formed to provide service to the regional research networks and provide alternate network access, UUCP-based email and Usenet News to the public. The first dialup on the West Coast, was Best Internet[22] - now Verio, opened in 1986 1986.. The The firs firstt dial dialup up ISP ISP in the the East East was was worl world. d.st std. d.co com, m, opened in 1989. This This caus caused ed cont contro rove vers rsy y amon amongs gstt univ univer ersi sity ty user users, s, who who were were outr outrag aged ed at the the idea idea of none nonedu duca cati tion onal al use use of thei theirr netw networ orks ks.. Eventually, it was the commercial Internet service providers who brought prices low enough that junior colleges and other schools
could afford to participate in the new arenas of education and research. By 1990, ARPANET had been overtaken and replaced by newer networking technologies and the project came to a close. In 1994, the NSFNet, now renamed ANSNET (Advanced Networks and Servic Services) es) and allowi allowing ng non-pr non-profi ofitt corpor corporati ations ons access access,, lost lost its stan standi ding ng as the the back backbo bone ne of the the Inte Intern rnet et.. Both Both gove govern rnme ment nt institutions and competing commercial providers created their own backbones and interconnections. Regional network access points (NAPs) became the primary interconnections between the many networks and the final commercial restrictions ended. IETF AND A STANDARD FOR STANDARDS
The Internet has developed a significant subculture dedicated to the idea that the Internet is not owned or controlled by any one per perso son, n, comp compan any, y, grou group, p, or orga organi niza zati tion on.. Neve Nevert rthe hele less ss,, some some standardization and control is necessary for the system to function.
The liberal Request for Comments (RFC) publication procedure engendered confusion about the Internet standardization process, and led to more formalization of official accepted standards. The IETF started in January of 1985 as a quarterly meeting of U.S. gove govern rnme ment nt fund funded ed rese resear arch cher ers. s. Repr Repres esen enta tati tive vess
from from nonnon-
government vendors were invited starting with the fourth IETF meeting in October of that year. Acceptance of an RFC by the RFC Editor for publication does not automatically make the RFC into a standard. It may be recognized as such such by the IETF IETF onl only aft after exper xperiiment mentat atiion, on, use use, and and acceptance have proved it to be worthy of that designation. Official standards are numbered with a prefix "STD" and a number, similar to the the RFC RFC nami naming ng styl style. e. Howe Howeve ver, r, even even afte afterr beco becomi ming ng a stan standa dard rd,, most most are are stil stilll comm common only ly refe referr rred ed to by thei theirr RFC RFC number. In 1992, the Internet Society, a professional membership society, was formed and the IETF was transferred to operation under it as an independent international standards body.
NIC, InterNIC, IANA and ICANN
The firs firstt cent centra rall auth author orit ity y to coor coordi dina nate te the the oper operat atio ion n of the the network was the Network Information Centre (NIC) at Stanford Rese Resear arch ch Inst Instit itut utee (SRI (SRI)) in Menl Menlo o Park Park,, Cali Califo forn rnia ia.. In 1972 1972,, mana manage geme ment nt of thes thesee issu issues es was was give given n to the the newl newly y crea create ted d Internet Assigned Numbers Authority (IANA). In addition to his role as the RFC Editor, Jon Postel worked as the manager of IANA until his death in 1998. As the early ARPANET grew, hosts were referred to by names, and a HOSTS.TXT file would be distributed from SRI International to each host on the network. As the network grew, this became cumbersome. A technical solution came in the form of the the Doma Domain in Name Name Syst System em,, crea create ted d by Paul Paul Mock Mockap apet etri ris. s. The The Defense Data Network—Network Information Center (DDN-NIC) at SRI handle handled d all all regist registrat ration ion servic services, es, includ including ing the top-le top-level vel domains (TLDs) of .mil, .gov, .edu, .org, .net, .com and .us, root nameserver administration and Internet number assignments under
a United States Department of Defense contract.[23] In 1991, the Defe Defens nsee Info Inform rmat atio ion n Syst System emss Agen Agency cy (DIS (DISA) A) awar awarde ded d the the administration and maintenance of DDN-NIC (managed by SRI up until this point) to Government Systems, Inc., who subcontracted it to the small private-sector Network Solutions, Inc. Since at this point in history most of the growth on the Internet was comi coming ng from from nonnon-mi mili lita tary ry sour source ces, s, it was deci decide ded d that that the the Department of Defense would no longer fund registration services outs outsid idee of the the .mil .mil TLD. TLD. In 1993 1993 the the U.S. U.S. Natio ationa nall Scie Scienc ncee Foundation, after a competitive bidding process in 1992, created the InterNIC to manage the allocations of addresses and management of the address databases, and awarded the contract to three organizations. Registration Services would be provided by Netwo Network rk Soluti Solutions ons;; Direct Directory ory and Databa Database se Servic Services es would would be provided by AT&T; and Information Services would be provided by General Atomics.
In 1998 1998 both both IANA IANA and and Inte InterN rNIC IC were were reor reorga gani nize zed d unde underr the the control of ICANN, a California non-profit corporation contracted by by the the US Depa Depart rtme ment nt of Comm Commer erce ce to mana manage ge a numb number er of Internet-related tasks. The role of operating the DNS system was pri priva vati tize zed d and and open opened ed up to comp compet etit itio ion, n, whil whilee the the cent centra rall management of name allocations would be awarded on a contract tender basis. USE AND CULTURE
E-mail and Usenet E-ma E-mail il is ofte often n call called ed the the kill killer er appl applic icat atio ion n of the the Inte Intern rnet et.. However, it actually predates the Internet and was a crucial tool in creating it. E-mail started in 1965 as a way for multiple users of a time-sharing mainframe computer to communicate. Although the history is unclear, among the first systems to have such a facility were SDC's Q32 and MIT's CTSS. The ARPANET computer network made a large contribution to the evolution of e-mail. There is one report indicating experimental inte inter-s r-sys yste tem m e-ma e-mail il tran transf sfer erss on it shor shortl tly y afte afterr ARPA ARPANE NET' T'ss
creation. In 1971 Ray Tomlinson created what was to become the stan standa dard rd Inte Intern rnet et e-ma e-mail il addr addres esss form format at,, usin using g the the @ sign sign to separate user names from host names. A number of protocols were developed to deliver e-mail among groups groups of time-s time-shar haring ing comput computers ers over over altern alternat ative ive transm transmiss ission ion systems, such as UUCP and IBM's VNET e-mail system. E-mail could be passed this way between a number of networks, including ARPANET, BITNET and NSFNet, as well as to hosts connected directly to other sites via UUCP. In addition, addition, UUCP allowed allowed the publicati publication on of text files files that could be read by many others. The News software developed by Steve Daniel and Tom Truscott in 1979 was used to distribute news and bulletin board-like messages. This quickly grew into discussion groups, known as newsgroups, on a wide range of topics. On ARPANET and NSFNet similar discussion groups would form via mailing lists, discussing both technical issues and more culturally focused topics (such as science fiction, discussed on the sflovers mailing list).
From gopher to the WWW
As the Internet grew through the 1980s and early 1990s, many people realized the increasing need to be able to find and organize files and information. Projects such as Gopher, WAIS, and the FTP Archive list attempted to create ways to organize distributed data. Unfortunately, these projects fell short in being able to accommodate all the existing data types and in being able to grow without bottlenecks.[citation needed]
One of the most promising user interface paradigms during this per perio iod d was was hype hypert rtex ext. t. The The tech techno nolo logy gy had had been been insp inspir ired ed by Vannevar Bush's "Memex" and developed through Ted Nelson's research on Project Xanadu and Douglas Engelbart's research on NLS NLS.. Many Many smal smalll self self-co -cont ntai aine ned d hype hypert rtex extt syst system emss had had been been create created d before before,, such such as Apple Apple Comput Computer' er'ss Hyper HyperCar Card. d. Gophe Gopher r became the first commonly-used hypertext interface to the Internet.
While Gopher menu items were examples of hypertext, they were not commonly perceived in that way. In 1989, whilst working at CERN, Tim Berners-Lee invented a netw networ orkk-ba base sed d impl implem emen enta tati tion on of the the hype hypert rtex extt conc concep ept. t. By releasing his invention to public use, he ensured the technology woul would d beco become me wide widesp spre read ad.. One One earl early y popu popula larr web web brow browse ser, r, modeled after HyperCard, was ViolaWWW. Scho Schola lars rs gene genera rall lly y agre agree, e,[c [cit itat atio ion n need needed ed]] howe howeve ver, r, that that the the turning point for the World Wide Web began with the introduction of the Mosaic web browser in 1993, a graphical browser developed by a team at the National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign (NCSA-UIUC), led by Marc Andreessen. Funding for Mosaic came from the HighPerformance Computing and Communications Initiative, a funding progr program am initia initiated ted by then-S then-Sena enator tor Al Gore's Gore's High High Perfor Performan mance ce Computing and Communication Act of 1991 also known as the Gore Bill . Indeed, Mosaic's graphical interface soon became more popular than Gopher, which at the time was primarily text-based,
and the WWW became the preferred interface for accessing the Internet. (Gore's reference to his role in "creating the Internet", however, was ridiculed in his presidential election campaign. See the full article Al Gore and information technology). Mosa Mosaic ic was was even eventu tual ally ly supe supers rsed eded ed in 1994 1994 by Andr Andree eess ssen en's 's Netscape Navigator, which replaced Mosaic as the world's most popular browser. While it held this title for some time, eventually competition from Internet Explorer and a variety of other browsers almost completely displaced it. Another important event held on Janu Januar ary y 11, 11, 1994 1994,, was was The The Supe Superh rhig ighw hway ay Summ Summit it at UCLA UCLA's 's Royce Hall. This was the "first public conference bringing together all of the major industry, government and academic leaders in the field [and] also began the national dialogue about the Information Superhighway and its implications." 24 Hours in Cyberspace, the "the largest one-day online event" (February 8, 1996) up to that date, took place on the then-active webs websit ite, e, cybe cyber2 r24. 4.co com. m. It was was head headed ed by phot photog ogra raph pher er Rick Rick Smolan.A
photographic
exhibition
was
unveiled
at
the
Smithsonian Institution's National Museum of American History on 23 January 1997, featuring 70 photos from the project.[40] Search engines
Even before the World Wide Web, there were search engines that atte attemp mpte ted d to orga organi nize ze the the Inte Intern rnet et.. The The firs firstt of thes thesee was was the the Archie search engine from McGill University in 1990, followed in 1991 by WAIS and Gopher. All three of those systems predated the invention of the World Wide Web but all continued to index the Web and the rest of the Internet for several years after the Web appeared. There are still Gopher servers as of 2006, although there are a great many more web servers. As the Web grew, search engines and Web directories were created to track pages on the Web and allow people to find things. The T he first full-t full-text ext We Web b search search engine engine was WebCra WebCrawle wlerr in 1994. 1994. Before Before WebCrawler, only Web page titles were searched. Another early search search engine, engine, Lycos, was created created in 1993 as a university university project, project, and was the first to achieve commercial success. During the late 1990s, both Web directories and Web search engines were popular
—Yahoo! (founded 1995) and Altavista (founded 1995) were the respective industry leaders. By August 2001, the directory model had begun to give way to search engines, tracking the rise of Google (founded 1998), which had developed new approaches to relevancy ranking. Directory features, while still commonly available, became after-thoughts to search engines. Databa Database se size, size, which which had been been a signif significa icant nt market marketing ing featur featuree through the early 2000s, was similarly displaced by emphasis on relevancy ranking, the methods by which search engines attempt to sort the best results first. Relevancy ranking first became a major issue circa 1996, when it became apparent that it was impractical to revi review ew full full list listss of resu result lts. s. Cons Conseq eque uent ntly ly,, algo algori rith thms ms for for rel releva evancy ncy
rank rankiing
have have
cont contin inu uousl ously y
impro mprov ved. ed.
Googl oogle' e'ss
PageRank method for ordering the results has received the most press, but all major search engines continually refine their ranking meth method odol olog ogie iess with with a view view towa toward rd impr improv ovin ing g the the orde orderi ring ng of results. As of 2006, search engine rankings are more important
than ever, so much so that an industry has developed ("search engine engine optimi optimizer zers", s", or "SEO") "SEO") to help help web-d web-deve evelop lopers ers improv improvee their search ranking, and an entire body of case law has developed around matters that affect search engine rankings, such as use of trad tradem emar arks ks in meta metata tags gs.. The The sale sale of sear search ch rank rankin ings gs by some some search engines has also created controversy among librarians and consumer advocates. Dot-com bubble
The suddenly low price of reaching millions worldwide, and the possibility of selling to or hearing from those people at the same moment when they were reached, promised to overturn established bus busin ines esss dogm dogmaa in adve advert rtis isin ing, g, mail mail-o -ord rder er sale sales, s, cust custom omer er relationship management, and many more areas. The web was a new new kill killer er app— app—it it coul could d brin bring g toge togeth ther er unre unrela late ted d buye buyers rs and and seller sellerss in seamle seamless ss and low-co low-cost st ways. ways. Vision Visionari aries es around around the world developed new business models, and ran to their nearest venture capitalist. Of course some of the new entrepreneurs were truly talented at business administration, sales, and growth; but the
majority were just people with ideas, and didn't manage the capital influx prudently. Additionally, many dot-com business plans were predicated on the assumption that by using the Internet, they would byp bypas asss the the dist distri ribu buti tion on chan channe nels ls of exis existi ting ng busi busine ness sses es and and therefore not have to compete with them; when the established bus busin ines esse sess with with stro strong ng exis existi ting ng bran brands ds deve develo lope ped d thei theirr own own Internet presence, these hopes were shattered, and the newcomers were left attempting to break into markets dominated by larger, more established businesses. Many did not have the ability to do so. The The dot-com bubble burst on March 10, 2000, when the technology heavy NASDAQ Composite index peaked at 5048.62 (intra-day peak 5132.52), more than double its value just a year before. By 2001, the bubble's deflation was running full speed. A majority of the dot-coms had ceased trading, after having burnt through their venture capital and IPO capital, often without ever making a profit.
Worldwide Online Population Forecast
In its "World "Worldwid widee Online Online Popula Populatio tion n Foreca Forecast, st, 2006 2006 to 2011," 2011," Jupi Jupite terR rRes esea earc rch h anti antici cipa pate tess that that a 38 perc percen entt incr increa ease se in the the number of people with online access will mean that, by 2011, 22 percent of the Earth's population will surf the Internet regularly. JupiterResearch says the worldwide online population will increase at a compound annual growth rate of 6.6 percent during the next five years, far outpacing the 1.1 percent compound annual growth rate for the planet's population as a whole. The report says 1.1 billion people currently enjoy regular access to the Web. North America will remain on top in terms of the number of people with with
onl online ine
acc access. ess. Acc Accordi ording ng to
Jupit upiteerRes rResea earc rch, h,
onli online ne
penetration rates on the continent will increase from the current 70 percent of the overall North American population to 76 percent by 2011. However, Internet adoption has "matured," and its adoption pace has slowed, in more developed countries including the United States, Canada, Japan and much of Western Europe, notes the report.
As the online population of the United States and Canada grows by about only 3 percent, explosive adoption rates in China and India will take place, says JupiterResearch. The report says China should reach an online penetration rate of 17 percent by 2011 and India should hit 7 percent during the same time frame. This growth is dire direct ctly ly rela relate ted d to infr infras astr truc uctu ture re deve develo lopm pmen entt and and incr increa ease sed d consumer purchasing power, notes JupiterResearch. By 2011, Asians will make up about 42 percent of the world's population with regular Internet access, 5 percent more than today, says the study. Pene Penetr trat atio ion n leve levels ls simi simila larr to Nort North h Amer Americ ica' a'ss are are foun found d in Scandi Scandinav navia ia and bigger bigger Wester Western n Europe European an nation nationss such such as the United Kingdom and Germany, but JupiterResearch says that a numb number er of Cent Centra rall Euro Europe pean an coun countr trie iess "are "are rela relati tive ve Inte Intern rnet et laggards." Brazil "with its soaring economy," is predicted by JupiterResearch to experience a 9 percent compound annual growth rate, the fastest
in Latin America, but China and India are likely to do the most to boost the world's online penetration in the near future. For the study, JupiterResearch defined "online users" as people who regularly access the Internet by "dedicated Internet access" devices. Those devices do not include cell phones.[41] Historiography
Some concerns have been raised over the historiography of the Inte Intern rnet et''s deve develo lopm pmeent. nt. Thi This is due due to lack of cent entral ralised sed documentation for much of the early developments that led to the Internet. "The Arpanet period is somewhat well documented because the corporation in charge - BBN - left a physical record. Moving into the the NSFN NSFNET ET era, era, it beca became me an extr extrao aord rdin inar aril ily y dece decent ntra rali lise sed d process. The record exists in people's basements, in closets. [...] So much of what happened was done verbally and on the basis of individual trust."
—Doug Gale
Cyberlaws Why Cyberlaws In India India became independent on 15th August, 1947. In the 49th year of Indian independence, Internet was commercially introduced in our country. The beginnings of Internet were extremely small and the growth of subscribers painfully slow. However as Internet has grown in our country, the need has been felt to enact the relevant Cyberlaws which are necessary to regulate Internet in India. This need for cyberlaws was propelled by numerous factors. Firstly, India has an extremely detailed and well-defined legal system in place. Numerous laws have been enacted and implemented and the foremost amongst them is The Constitution of India. We have interalia, amongst others, the Indian Penal Code, the Indian Evidence Act 1872, the Banker's Book Evidence Act, 1891 and the Reserve Bank of India Act, 1934, the Companies Act, and so on. However the arrival of Internet signalled the beginning of the rise of new and complex legal issues. It may be pertinent to mention that all the existing laws in place in India were enacted way back keeping in mind the relevant political, social, economic, and cultural scenario of that relevant time. Nobody then could really visualize about the Internet. Despite the brilliant acumen of our master draftsmen, the requirements of cyberspace could hardly ever be anticipated. As such, the coming of the Internet led to the emergence of numerous ticklish legal issues and problems which necessitated the enactment of Cyberlaws. Secondly, the existing laws of India, even e ven with the most benevolent and liberal interpretation, could not be interpreted in the light of the emerging cyberspace, to include all aspects relating to different activities in cyberspace. In fact, the practical experience and the wisdom of judgment found that it shall not be without major perils and pitfalls, if the existing laws were to be interpreted in the scenario of emerging cyberspace, without enacting new cyberlaws. As such, the need for enactment of relevant cyberlaws. Thirdly, none of the existing laws gave any legal validity or sanction to the activities in Cyberspace. For example, the Net is used by a large majority of users for email. Yet till today, email is not "legal" in our country. There is no law in the country, which gives legal validity, and sanction to email. Courts and
judiciary in our country have been reluctant to grant judicial recognition to the legality of email in the absence of any specific law having been enacted by the Parliament. As such the need has arisen for Cyberlaw. Fourthly, Internet requires an enabling and supportive legal infrastructure in tune with the times. This legal infrastructure can only be given by the enactment of the relevant Cyberlaws as the traditional laws have failed to grant the same. E-commerce, the biggest future of Internet, can only be possible if necessary legal infrastructure compliments the same to enable its vibrant growth. All these and other varied considerations created the conducive atmosphere for the need for enacting relevant cyberlaws in India. The Government of India responded by coming up with the draft of the first Cyberlaw of India - The Information Technology Bill, 1999. One question that is often asked is why should we have Cyberlaw in India, when a large chunk of the Indian population is below the poverty line and is residing in rural areas ? More than anything else, India, by its sheer numbers, as also by virtue of its extremely talented and ever growing IT population, is likely to become a very important Internet market in the future and it is important that we legislate Cyberlaws in India to provide for a sound legal and technical frame work which, in turn, could be a catalyst for growth and success of the Internet Revolution in India.
SUPPORTIVE CYBER LAW • Existing Statutes 1. Communica Communications tions and and Multimedi Multimediaa Act 1998(CMA) 1998(CMA) 2. Malaysian Communications and Multimedia Commission Act 1998 3. Digital Signature Act 1997 4. Computer Crimes Act 1997 5. Copyright Act (Amendment) Act 1997 6. Telemedicine Act 1997 7.
Optical Discs Act 2000
• Amendments of Statutes 1. Communications and Multimedia (Amendment) Bill 2004 2.
Communications
and
Multime imedia
(Amendment) Bill 2004. • Proposed Statutes 1. Personal Data Protection Act 2. Electronic Transactions Act (ETA) 3. E-Government Activities Act (EGA) 4. New Subsidiary Legislations
Commissi ssion
COMMUNICATIONS AND MULTIMEDIA ACT 1998
MAIN MAINFEATURES FEATURES Transparent Pro-Competition
The “mother” cyber law that provides for legislative, regulatory and institutional framework to cater for the convergence of the telecommunications, telecommunications, broadcasting and computing industries.
Less Regulation Flexible and Generic Emphasize Process Rather than Content Industry Self-Disci Self-Discipline pline Self-Discip line Regulatory Forebearance
REGULATORY FRAMEWORK (NEW LICENSING STRUCTURE) •
LICENCE o
INDIVIDUAL / CLASS
Network Facilities
Network Services
Content Application Services
Application Services
LICENSES ISSUED UNDER ACT 588 LICENSE Network Facilities Provider (NFP)
INDIVIDUAL
31
CLASS
24
Network Service Provider (NSP)
30
24
Application Service Provider (ASP)
80
95
Content Application Service Provider (CASP)
20
-
161
143
TOTAL
NEW AND MIGRATION LICENSES UNDER ACT 588 (INDIVIDUAL LICENSES)
LICENSES
MIGRATION
NEW
TOTAL
Network Facilities Provider
20
11
31
(NFP) Network Service Provider
19
11
30
(NSP) Application Service Provider
16
64
80
(ASP) Content Application Service
19
1
20
Provider (CASP) Total
74
87
161
DEVELOPMENT SINCE ACT 588
1. VISIBL VISIBLE E INCREA INCREASE SE IN CELLU CELLULAR LAR PENETRAT PENETRATION ION -From 12% or 2.7 million subscribers in 1999 to 43.6% or 11 million subscribers in 2003* 2. INCREASE INCREASE IN INTER INTERNET NET USERS USERS -
From 2.0 million in 1999 to 8.7 million iin n 2003*
3. MORE MORE CHOI CHOICE CES S FOR FOR CONS CONSUM UMER ERS S AND AND LO LOWE WER R COSTS OF SERVICES
- Streamyx service reduced by 30% - Lower charges for mobile services - More “free to air” TV stations – Channel 9, 8TV
INSTITUTIONAL FRAMEWORK
MINISTER •
MCMC INDUSTRY FORUMS
•
MECM TRIBUNAL
COMMUNICATIONS AND MULTIMEDIA COMMISSION ACT 1998 Power to establish an independent body to:
1.
Enfor force legi legisl slat atiion (CM (CMA 199 1998)
2.
Regulate industry
3.
Promote Ind Industry Dev Develo elopment
4.
Promo omote Indu Indus stry try Sel Selff-Re -Regul gulatio ation n
DIGITAL SIGNATURE ACT 1997 An Act to legalise digital signature
Facili Facilitat tate e e-com e-commer merce ce and and secu secure re on- line line transaction through the use of digital signatures Establishment of Certification Authority as the body responsible in issuing PKI, Private key, warranties and liabilities.
COMPUTER CRIME ACT 1997
The Act provides for: protection to companies, government and individuals from computer crimes in the digital era; clear definitions on criminal activities related to use of computers computers such such as cyber fraud, illegal access, interceptions, and illegal use of computers.
COMPUTER CRIME ACT 1997 The Act provides for: protection to companies, government and individuals from computer crimes in the digital era; clear definitions on criminal activities related to use of computers computers such such as cyber fraud, illegal access, interceptions, and illegal use of computers.
COMPUTER CRIME ACT 1997 Under-reporting Under-reporting of cyber crimes:
Maintaining their business and making profit;
Unwillingness Unwillingness to go through the legal process;
Expose confidential business information;
No pr prov ovis isio ion n for for vict victim im to re rece ceiv ivee re rest stit itut utio ion n for for th thee damage suffered.
COPYRIGHT ACT(A) 1997 Provides protection for multimedia works. Reflects up-to-date developments in copy rights issue. Clarify Clarify legal issues issues in digital digital transmission, transmission, use of multimedi multimediaa and its components.
TELEMEDICNE ACT(A) 1997 Provisions to regulate telemedicine activities:
Registration for practitioners;
Telemedicine Telemedicine practices by foreign practitioners; and
Medical
data
prescription
management
and
electronic
CELLULAR PENETRATION BETWEEN SELECTED COUNTRIES 2002 90
84.49 79.14
80
67.95
70
62.11
60 48.81
50
37.3
40
26.04
30
16.09
20 10
1.22
0 UK
Singapor e Kor ea ea, Rep
Japan
USA
Malay si sia
Thailand
China
India
Source: ITU@2003
COMPUTER AND INTERNET PENETRATION Computer Ownership - 4.2 Juta (16.7 %) Internet Penetration - 2.9 Juta (11.4 %)
18.0
16.7
16.0
14.5
14.0
12.5 11.4
12.0 %
10.5 9.4
10.0 8.0
7.9
8.8 7.1
6.1
6.0 4.0
2.9 1.8
2.0 1998
1999
2000 PCs
2001
2002
2003
Intern Internet et subscribers subscribers
Sumber : Sumber : MCMC
INTERNET PENETRATION BETWEEN SELECTED COUNTRIES 2002 60
55.2
54.0
53.8
50
44.9 40.6
40 31.6 30 20 7.8
10
4.6
1.6
0 Kor ea ea, Rep
Sin ga gapor e
USA
Japan
UK
Malay si sia
Thailand
China
India
Source: ITU@2003
BROADBAND PENETRATION RATES (%) AMONG SELECTED ASIAN COUNTRIES IN 2002 25
20
14.5m
15 465k
10 640k
5
196k
0
S o u th K Ho n g K %
1 9 .2 9
1 3 .3
T a iw a S in g a p 9 .1 5
6 .1 3
C h i n a M A L 1.5m A Y T h a i l a n19kI n d i a 0 .1 2
0 .0 8
0 .0 5
0 .0 2
Source:
Frost & Sullivan
AMENDMENT OF STATUTES. TH E
COMMUNICATIONS
AND
MULTIMEDIA
(AMENDMENT) BILL 2003 PURPOSE OF AMENDMENTS: AMENDMENTS:
TO
INSERT
THE
NECESSARY
SUBSTANTIVE
PROVIS PROVISION IONS S FOR THE EST ESTABL ABLISH ISHMEN MENT T OF AN INDEPENDENT APPEAL TRIBUNAL; AND
TO STRENGHTHEN THE CURRENT REGULATORY AND LICENSING REGIME.
AMENDMEN AMENDMENTS TS - SUBSIDIA SUBSIDIARY RY LEGISLATIO LEGISLATIONS NS UNDER UNDER THE COMMUNICATIONS AND MULTIMEDIA ACT 1998
Communications and Multimedia (Licensing) Regulations 2000;
Communications and Multimedia (Spectrum) Regulations 2000;
Comm Commun unic icat atio ions ns
and and
Mult Multim imed edia ia
(Tec (T echn hnic ical al
Stan Standa dard rdss)
Regulations 2000;
Communications and Multimedia (Spectrum) (Exemption) Order 2000;
Communications and Multimedia (Licensing) (Exemption) Order 2000
Communications and Multimedia (USP) Regulations 2002
Communications and Multimedia (Rates) Rules 2002
Notification of Issuance of Class Assignments
ENACTMENT OF NEW LAWS To pr provi ovide de legal legal cer certai tainty nty for e-t e-tran ransac sactio tions ns und under ertak taken en by bussines bu inesse sess or Gove Goverrnme nment nt,, two ne new w legisl gislat atiions ons will will be introduced:-
- Electronic Transactions Bill – to address electronic transactions and communications. communications. E-Go Gove vern rnm men entt Acti Activi viti ties es Bill Bill – to supp suppor ortt and and - Epromote electronic government.
PRIVACY AND INFORMATION SECURITY
ENSURING ON-LINE TRUST AND CONFIDENCE
Two aspects related to on-line trust and confidence: Privacy and personal data protection (PDP); and Security of electronic transactions.
PERSONAL DATA PROTECTION BILL
Promote secured electronic environment
PDP Protect personal data
Enhance consumer trust and confidence
Encourage electronic transactions
Privacy is a Shared Responsibility Roles to play for individuals, individuals , industry, and government. Individuals should be able to make informed choices and be protected from harm & fraud Industry should ensure fair information practices. In some areas, Governments Governments must choose whether to limit individual individual control over data to achieve larger societal benefits (e.g. security, health etc.). A balanced approach enables individuals to benefit from responsible commercial uses of personal information
WEB APPLICATION/WRITING WEB PROJECTS/ WEB OBJECTS/ WEB USERS
In software engineering, engineering, a Web application is an application that is accessed via Web browser over browser over a network such as the Internet or an intranet intranet.. It is also a computer software application that is coded in a browser-supported language (such as HTML HTML,, JavaScript JavaScript,, Java,, etc.) and reliant on a common Java co mmon web browser to render the applicationexecutable application executable.. Web applications are popular due to the ubiquity of a client client,, sometimes called a thin client.. The ability to update and maintain Web applications without distributing and client installing software on potentially thousands of client computers is a key reason for their popularity. Common Web applications include Webmail Webmail,, online retail sales, sales, online auctions,, wikis auctions wikis,, discussion boards, boards, Weblogs Weblogs,, MMORPGs and many other functions.
History In earlier types of client-server of client-server computing, computing, each application had its own client program which served as its user interface and had to be separately installed on each user's personal computer . An upgrade to the server part of the application would typically require an upgrade to the clients installed on each user workstation, adding to the support cost and decreasing productivity decreasing productivity.. In contrast, Web applications dynamically generate a series of Web of Web documents in a standard format supported by common browsers such as HTML HTML//XHTML XHTML.. Client-side scripting in a standard language such as JavaScript is commonly included to add dynamic elements to the user interface. Generally, each individual Web page is delivered to the client as a static document, but the sequence of pages can provide an interactive
experience, as user input is returned through Web form elements embedded in the page markup. During the session, the Web browser interprets and displays the pages, and acts as the universal client for any Web application.
Interface
Webconverger operating Webconverger operating system provides system provides an interface for web applications. The Web interface places very few limits on client functionality. Through Java Java,, JavaScript,, DHTML JavaScript DHTML,, Flash and other technologies, application-specific methods such as drawing on the screen, playing audio, and access to the keyboard and mouse are all possible. Many services have worked to combine all of these into a more familiar interface that adopts the appearance of an operating system. General purpose techniques such as drag and drop are also supported by these technologies. Web developers often use client-side scripting to add functionality, especially to create an interactive experience that does not require page reloading (which many users find disruptive)[citation needed ]. Recently, technologies have been developed to coordinate client-side scripting with server-side technologies such as PHP PHP.. Ajax Ajax,, a web development technique using a combination of various technologies, is an example of technology which creates a more interactive experience.
Technical considerations A significant advantage of building Web applications to support standard browser features is that they should perform as specified regardless of the op erating system or OS version installed on a given client. c lient. Rather than creating clients for MS for MS Windows, Windows, Mac OS X, GNU/Linux GNU/Linux,, and other operating other operating systems, systems, the application can be written once and deployed almost anywhere. However, inconsistent implementations of the HTML HTML,, CSS CSS,, DOM and other browser specifications can cause problems in web application development and support. Additionally, the ability of users to customize many of the display settings of their browser (such as selecting different font sizes, colors, and typefaces, or disabling scripting support) can interfere with consistent implementation of a Web application. Another approach is to use Adobe Flash or Java or Java applets to provide some or all of the user interface. Since most Web browsers include support for these technologies (usually through plug-ins), Flash- or Java-based applications can be implemented with much of the same ease of deployment. Because Becau se they allow the programmer greater control over the interface, they bypass many browser-configuration issues, although incompatibilities between Java or Flash implementations on the client can introduce different complications. Because of their architectural similarities to traditional client-server applications, with a somewhat "thick" client, there is some dispute over whether to call
systems of this sort "Web applications"; an alternative term is "Rich "Rich Internet Application" Application" (RIA).
Structure Though many variations are possible, a Web application is commonly structured as a three-tiered application. In its most common form, a Web browser is the first tier, an engine using some dynamic Web content technology (such as ASP ASP,, ASP.NET ASP.NET,, CGI CGI,, ColdFusion,, JSP/Java ColdFusion JSP/Java,, PHP PHP,,embPerl embPerl,, Python Python,, or Ruby or Ruby on Rails) Rails) is the middle tier, and a database is the third tier. The Web browser sends requests to the middle tier, which services them by making queries and updates against the database and generates a user interface. But there are some who view a web application as a Two-Tier architecture.
Business use An emerging strategy for application software companies is to p rovide Web access to software previously distributed as local applications. Depending on the type of application, it may require the development of an entirely different browser-based interface, or merely adapting an existing application to use different presentation technology. These programs allow the user to pay a monthly or yearly fee for use of a software application without having to install it on a local hard drive. A company which follows this strategy is known as an application service provider (ASP), provider (ASP), and ASPs are currently receiving much attention in the software industry.
Writing Web applications There are many Web application frameworks which facilitate rapid application development by allowing the programmer to define a high-level description of the program. In addition, there is potential for the development of applications on Internet operating systems, systems, although currently there are not many viable platforms that fit this model. The use of Web application frameworks can often reduce the number of errors in a program, both by making the code more simple, and by allowing one team to concentrate just on the framework. In applications which are exposed to constant hacking attempts on the Internet, security-related problems caused by errors in the program are a big issue. Frameworks may also promote the use of best practices such as GET after POST. POST.
Web Application Security The Web Application Security Consortium (WASC) and OWASP are projects developed with the intention of documenting how to avoid security problems in Web applications. A
Web Application Security Scanner is Scanner is specialized software for detecting security problems in web applications.
Applications
Wikipedia application running in Mozilla Firefox. Firefox. Browser applications typically include simple office software (word processors, spreadsheets, and presentation tools) and can also include more advanced application such as project management software, CAD Design Software, and point-of-sale applications. Examples
Word processor and Spreadsheet: Google Docs & Spreadsheets CRM Software: SalesForce.com •
•
Benefits Browser Applications typically require little or no disk space, upgrade au tomatically with new features, integrate easily into other web p rocedures, such as email and searching. They also provide cross-platform compatibility (i.e Mac or Windows) because they operate within a web browser window.
Disadvantages Standards compliance is an issue with any non-typical office d ocument creator, which causes problems when file sharing and collaboration becomes critical. Also, Browser Applications rely on application files accessed on remote servers through the internet. Therefore, when connection is interrupted, the application is no longer usable. Google Gears is a beta platform to combat this issue and improve the usability of Browser Applications.
As the Internet grew into a major maj or player on the global economic front, so did the number of investors who were interested in its development. So, you may wonder, how does the Internet continue to play a major role in communications, media and news? The key words are: Web Application Projects. Web applications are business strategies and policies implemented on the Web through the use of User, Business and Data services. These tools are where the future lies. li es. In this article, I'll take you through the essential phases in the life cycle of a Web application project, explain what options you have, and help you formulate a plan for successful Web application endeavors of your own. First, though, let's take tak e a brief overview of Web applications. Who Needs Web Applications and Why? There are many entities that require applications for the Webone example would be Business-to-Business interaction. Many companies in the world today demand to do business with each other over secure and private networks. This process is becoming increasingly popular with a lot of overseas companies who outsource projects to each other. From the simple process p rocess of transferring funds into a bank account, to deploying a large scale Web services network that updates pricing information globally, the adoption of a Web applications infrastructure is vital for many businesses. The Web Application Model
The Web application model, like many software development models, is constructed upon 3 tiers: User Services, Business Services and Data Services. This model breaks an application into a network of consumers and suppliers of services. The User Service tier creates a visual gateway for the consumer to interact with the application. This can range from basic HTML and DHTML to complex COM components and Java applets. The user services then grab business logic and procedures from the Business Services. This tier can range from Web scripting in ASP / /PHP PHP / /JSP JSP to server side programming such as TCL, CORBA and PERL PERL,, that allows the user to perform complex actions through a Web interface. The final tier is the Data Service layer. Data services store, retrieve and update information at a high level. Databases, file systems, and writeable media are all examples of Data storage and retrieval devices. For Web applications, however, databases are most practical. Databases allow developers to store, retrieve, add to, and update categorical information in a systematic and organized fashion. Choosing the Right Project
Choosing the right types of projects to work on is an extremely important part of the Web application development plan. pl an. Assessing your resources, technical skills, and publishing capabilities should be your first goal. Taking the 3 tiers into consideration, devise a list of all available resources that can be categorically assigned to each tier. The next consideration should be the cost. Do you y ou have a budget with which to complete this project? How much will it cost you to design, develop and deliver a complete project with a fair amount of success? These are questions that should be answered before you sign any deals or contracts. Let's look at an example. A company called ABC needs to develop a Web application that will display sales information created by different sales agents. The data is updated daily through a completely automated process from all 3 service tiers. The client tells you that this entire project must be done in ASP/SQL server and that you should host the application as well. After assessing all your resources, you and your team come to a conclusion that the company is unable to do data backups on a daily basis. After further discussion, you realize that this is a very important part of the setup for your y our client, and you should not risk taking a chance with the project. It's very likely that you will be more prepared next time around, when a similar project lands on your desk, so you decline the job and recommend someone else who has the capabilities to do it right now. The Phases in a Web Application Project The Web application development process has 4 phases:
1. Envi Envisio sionin ning g the nat nature ure and dire directio ction n of the the projec projectt 2. De Devi visi sing ng th the e pl plan an 3. Development 4. Te Test sting ing,, sup suppo port rt an and d sta stabil bilit ity y Let's look at each of these in more detail. 1. Envisioning the nature and direction of the project In this phase, the management and developers assigned to the project come together and establish the goals that the solution must achieve. This includes recognizing the limitations that are placed on the project, scheduling, and versioning of the application. By the end of this phase, there should be clear documentation on what the application will achieve. 2. Devising the plan In this phase, you and your team must determine the "how's" of the application. What scripting language is most appropriate, which features must be included, and how long will it take? These are some of the questions that must be answered through this planning phase. The main tangents at this point are the project plan and functional specification. The project plan determines a timeframe of events and tasks, while the functional specification outlines in detail how the application will function and flow. 3. Development
Once the project plan and functional specification are ready, a baseline is set for the development work to begin. The programmer/s or Web developer/s begin coding, testing and publishing data. This phase establishes the data variables, entities and coding procedures that will be used throughout the remainder of the project. A milestone m ilestone document is prepared by the development team, which is then handed to management for review. 4. Testing, support and stability The stability phase of the application project mainly focuses on testing and the removal of bugs, discrepancies and network issues that may otherwise cause the application to fail. It is here that policies and procedures are established for a successful support system. Planning for a Successful Web Development Project In order to drastically minimize the risk of project failure, I've always approached my application development projects in the following sequence. 1. Identify business logic and entities Start by gathering information on everything you have. If you are going to be working with databases, begin by enumerating how many entities will be used in the business logic. For example, if your program implements sales data, a sales ticket would be an entity. Once you've identified all your entities, establish a clear guideline for their relationships. This can be done via presentations, flowcharts or even reports.
2. Create a functional specification and project plan This part, in my opinion, is i s the most important part of the project. Functional specifications (or functional specs) are a map, or blueprint for how you want a particular Web application to look and work. The spec details what the finished product will do, user interaction, and its look and feel. An advantage of writing a functional spec is that it streamlines the development process. It takes discrepancies and guesswork out of the programming process, because the level of detail that goes into the plan makes it possible to minimize the misunderstanding that's usually associated with project mishaps. See examples of well written functional specs at RayComm.com RayComm.com.. Once the functional spec is finished, a project plan must be devised. A project plan is a timeline of tasks and events that will take place during the project. The project or program manager is normally the person who creates a project plan, p lan, and their primary focus is to detail task notes while being able to accommodate scheduling and resource information. You can download a sample Excel file for a project plan at Method123.com.. Method123.com 3. Bring the application model into play As discussed earlier, the application model consists of 3 tiers The User, Business and Data service tiers, each of which serves a substantial purpose. Practically speaking, it's always best to start with the data tier, because you've already identified your entities and understand their relationships. The data tier can be an SQL server database, a text file, or even the powerful and robust Oracle. Create tables, relationships, jobs, and procedures depending on what
platform you have chosen. If the data is a warehouse (i.e. the data already exists and does not depend on real time interaction), then make sure that new and additional data can be added securely and in a scalable fashion. A quick tip: using views in SQL server/Oracle can improve dramatically the productivity and performance of your application. They increase speed because they are "stored queries" that don't have a physical existence. The Business services tier, in my opinion, is the heart of the application. It involves the implementation of business logic into the scripting or programming language. At this stage, make sure you've already set up your environment for testing and debugging. Always test on at least 2 instances in your application, after all, what may work perfectly for you, may not do so well on other platforms or machines. ASP, XML XML,, PHP, JSP and CGI are some examples of server side scripting languages used at the business service level. Whichever language you choose, make sure that it's capable of handling all the business logic presented in the functional specification. The last is the user tier, which is absolutely vital for the interactive and strategic elements in the application. It provides the user with a visual gateway to the business service by placing images, icons, graphics and layout elements in strategic areas of interest, most commonly, based on management research. If you'll be developing the user tier yourself, be sure to have studied your competition. The last thing you need is for your application to look exactly the same as someone else's. 4. Develop a support scheme
Being able to support and stabilize your application is very important. Define a procedure call for cases of failure, mishaps or even downtime. Give your customers the ability to contact you in the case of an emergency relating to the program. A good example of a support scheme is a ticket tracking system. This system allows users to file fil e cases pertaining to a support request and the support team, then makes the case track able. This means that the request is identifiable by a unique code or number. Although ticket-tracking systems are normally used by hosting companies or large scale ASP's (Application Service Providers), they still serve a valuable purpose in helping keep the application stable.