In the context of the World Wide Web, a bookmark is a Uniform Resource Identifier, stored for retrieval in any of various storage formats. All modern web browsers include bookmark features. Bookmarks are called favorites or Internet shortcuts in Internet Explorer, by virtue of that browser's large market share, these terms have been synonymous with bookmark since the first browser war. Bookmarks are accessed through a menu in the user's web browser, folders are used for organization. In addition to bookmarking methods within most browsers, many external applications offer bookmark management. Bookmarks have been incorporated in browsers since the Mosaic browser in 1993. Bookmark lists were called Hotlists in previous versions of Opera. Other early web browsers such as ViolaWWW and Cello had bookmarking features. With the advent of social bookmarking, shared bookmarks have become a means for users sharing similar interests to pool web resources, or to store their bookmarks in such a way that they are not tied to one specific computer or browser.
Live bookmarks are updated automatically. Comparison of browser synchronizers Enterprise bookmarking Favicon Smart keyword Social bookmark link generator Social bookmarking XBEL Bookmark Managers at Curlie
Cascading Style Sheets
This cascading priority scheme is predictable. The CSS specifications are maintained by the World Wide Web Consortium. Internet media type text/css is registered for use with CSS by RFC 2318; the W3C operates a free CSS validation service for CSS documents. In addition to HTML, other markup languages support the use of CSS including XHTML, plain XML, SVG, XUL. CSS has a simple syntax and uses a number of English keywords to specify the names of various style properties. A style sheet consists of a list of rules; each rule or rule-set consists of one or more selectors, a declaration block. In CSS, selectors declare which part of the markup a style applies to by matching tags and attributes in the markup itself. Selectors may apply to the following: all elements of a specific type, e.g. the second-level headers h2 elements specified by attribute, in particular: id: an identifier unique within the document class: an identifier that can annotate multiple elements in a document elements depending on how they are placed relative to others in the document tree.
Classes and IDs are case-sensitive, start with letters, can include alphanumeric characters and underscores. A class may apply to any number of instances of any elements. An ID may only be applied to a single element. Pseudo-classes are used in CSS selectors to permit formatting based on information, not contained in the document tree. One example of a used pseudo-class is:hover, which identifies content only when the user “points to” the visible element by holding the mouse cursor over it, it is #elementid: hover. A pseudo-class classifies document elements, such as:link or:visited, whereas a pseudo-element makes a selection that may consist of partial elements, such as::first-line or::first-letter. Selectors may be combined in many ways to achieve great flexibility. Multiple selectors may be joined in a spaced list to specify elements by location, element type, id, class, or any combination thereof; the order of the selectors is important. For example, div.myClass applies to all elements of class myClass that are inside div elements, whereas.myClass div applies to all div elements that are in elements of class myClass.
The following table provides a summary of selector syntax indicating usage and the version of CSS that introduced it. A declaration block consists of a list of declarations in braces; each declaration itself consists of a property, a colon, a value. If there are multiple declarations in a block, a semi-colon must be inserted to separate each declaration. Properties are specified in the CSS standard; each property has a set of possible values. Some properties can affect any type of element, others apply only to particular groups of elements. Values may be keywords, such as "center" or "inherit", or numerical values, such as 200px, 50vw or 80%. Color values can be specified with keywords, hexadecimal values, RGB values on a 0 to 255 scale, RGBA values that specify both color and alpha transparency, or HSL or HSLA values. Before CSS, nearly all presentational attributes of HTML documents were contained within the HTML markup. All font colors, background styles, element alignments and sizes had to be explicitly described repeatedly, within the HTML.
CSS lets authors move much of that information to another file, the style sheet, resulting in simpler HTML. For example, sub-headings, sub-sub-headings, etc. are defined structurally using HTML. In print and on the screen, choice of font, size and emphasis for these elements is presentational. Before CSS, document authors who wanted to assign such typographic characteristics to, all h2 headings had to repeat HTML presentational markup for each occurrence of that heading type; this made documents more complex and more error-prone and difficult to maintain. CSS allows the separation of presentation from structure. CSS can define color, text alignment, borders, spacing and many other typographic characteristics, can do so independently for on-screen and printed views. CSS defines non-visual styles, such as reading speed and emphasis for aural text readers; the W3C has now deprecated the use of all presentational HTML markup. For example, under pre-CSS HTML, a heading element defined with red text would be written as: Using CSS, the sam
Open-source software is a type of computer software in which source code is released under a license in which the copyright holder grants users the rights to study and distribute the software to anyone and for any purpose. Open-source software may be developed in a collaborative public manner. Open-source software is a prominent example of open collaboration. Open-source software development generates an more diverse scope of design perspective than any company is capable of developing and sustaining long term. A 2008 report by the Standish Group stated that adoption of open-source software models have resulted in savings of about $60 billion per year for consumers. In the early days of computing and developers shared software in order to learn from each other and evolve the field of computing; the open-source notion moved to the way side of commercialization of software in the years 1970-1980. However, academics still developed software collaboratively. For example Donald Knuth in 1979 with the TeX typesetting system or Richard Stallman in 1983 with the GNU operating system.
In 1997, Eric Raymond published The Cathedral and the Bazaar, a reflective analysis of the hacker community and free-software principles. The paper received significant attention in early 1998, was one factor in motivating Netscape Communications Corporation to release their popular Netscape Communicator Internet suite as free software; this source code subsequently became the basis behind SeaMonkey, Mozilla Firefox and KompoZer. Netscape's act prompted Raymond and others to look into how to bring the Free Software Foundation's free software ideas and perceived benefits to the commercial software industry, they concluded that FSF's social activism was not appealing to companies like Netscape, looked for a way to rebrand the free software movement to emphasize the business potential of sharing and collaborating on software source code. The new term they chose was "open source", soon adopted by Bruce Perens, publisher Tim O'Reilly, Linus Torvalds, others; the Open Source Initiative was founded in February 1998 to encourage use of the new term and evangelize open-source principles.
While the Open Source Initiative sought to encourage the use of the new term and evangelize the principles it adhered to, commercial software vendors found themselves threatened by the concept of distributed software and universal access to an application's source code. A Microsoft executive publicly stated in 2001 that "open source is an intellectual property destroyer. I can't imagine something that could be worse than this for the software business and the intellectual-property business." However, while Free and open-source software has played a role outside of the mainstream of private software development, companies as large as Microsoft have begun to develop official open-source presences on the Internet. IBM, Oracle and State Farm are just a few of the companies with a serious public stake in today's competitive open-source market. There has been a significant shift in the corporate philosophy concerning the development of FOSS; the free-software movement was launched in 1983. In 1998, a group of individuals advocated that the term free software should be replaced by open-source software as an expression, less ambiguous and more comfortable for the corporate world.
Software licenses grant rights to users which would otherwise be reserved by copyright law to the copyright holder. Several open-source software licenses have qualified within the boundaries of the Open Source Definition; the most prominent and popular example is the GNU General Public License, which "allows free distribution under the condition that further developments and applications are put under the same licence", thus free. The open source label came out of a strategy session held on April 7, 1998 in Palo Alto in reaction to Netscape's January 1998 announcement of a source code release for Navigator. A group of individuals at the session included Tim O'Reilly, Linus Torvalds, Tom Paquin, Jamie Zawinski, Larry Wall, Brian Behlendorf, Sameer Parekh, Eric Allman, Greg Olson, Paul Vixie, John Ousterhout, Guido van Rossum, Philip Zimmermann, John Gilmore and Eric S. Raymond, they used the opportunity before the release of Navigator's source code to clarify a potential confusion caused by the ambiguity of the word "free" in English.
Many people claimed that the birth of the Internet, since 1969, started the open-source movement, while others do not distinguish between open-source and free software movements. The Free Software Foun
Hypertext Transfer Protocol Secure is an extension of the Hypertext Transfer Protocol. It is used for secure communication over a computer network, is used on the Internet. In HTTPS, the communication protocol is encrypted using Transport Layer Security, or its predecessor, Secure Sockets Layer; the protocol is therefore often referred to as HTTP over TLS, or HTTP over SSL. The principal motivation for HTTPS is authentication of the accessed website and protection of the privacy and integrity of the exchanged data while in transit, it protects against man-in-the-middle attacks. The bidirectional encryption of communications between a client and server protects against eavesdropping and tampering of the communication. In practice, this provides a reasonable assurance that one is communicating without interference by attackers with the website that one intended to communicate with, as opposed to an impostor. HTTPS connections were used for payment transactions on the World Wide Web, e-mail and for sensitive transactions in corporate information systems.
Since 2018, HTTPS is used more by web users than the original non-secure HTTP to protect page authenticity on all types of websites. The Uniform Resource Identifier scheme HTTPS has identical usage syntax to the HTTP scheme. However, HTTPS signals the browser to use an added encryption layer of SSL/TLS to protect the traffic. SSL/TLS is suited for HTTP, since it can provide some protection if only one side of the communication is authenticated; this is the case with HTTP transactions over the Internet, where only the server is authenticated. HTTPS creates a secure channel over an insecure network; this ensures reasonable protection from eavesdroppers and man-in-the-middle attacks, provided that adequate cipher suites are used and that the server certificate is verified and trusted. Because HTTPS piggybacks HTTP on top of TLS, the entirety of the underlying HTTP protocol can be encrypted; this includes the request URL, query parameters and cookies. However, because host addresses and port numbers are part of the underlying TCP/IP protocols, HTTPS cannot protect their disclosure.
In practice this means that on a configured web server, eavesdroppers can infer the IP address and port number of the web server that one is communicating with, as well as the amount and duration of the communication, though not the content of the communication. Web browsers know how to trust HTTPS websites based on certificate authorities that come pre-installed in their software. Certificate authorities are in this way being trusted by web browser creators to provide valid certificates. Therefore, a user should trust an HTTPS connection to a website if and only if all of the following are true: The user trusts that the browser software implements HTTPS with pre-installed certificate authorities; the user trusts the certificate authority to vouch only for legitimate websites. The website provides a valid certificate; the certificate identifies the website. The user trusts. HTTPS is important over insecure networks, as anyone on the same local network can packet-sniff and discover sensitive information not protected by HTTPS.
Additionally, many free to use and paid WLAN networks engage in packet injection in order to serve their own ads on webpages. However, this can be exploited maliciously in many ways, such as injecting malware onto webpages and stealing users' private information. HTTPS is very important for connections over the Tor anonymity network, as malicious Tor nodes can damage or alter the contents passing through them in an insecure fashion and inject malware into the connection; this is one reason why the Electronic Frontier Foundation and the Tor project started the development of HTTPS Everywhere, included in the Tor Browser Bundle. As more information is revealed about global mass surveillance and criminals stealing personal information, the use of HTTPS security on all websites is becoming important regardless of the type of Internet connection being used. While metadata about individual pages that a user visits is not sensitive, when combined, they can reveal a lot about the user and compromise the user's privacy.
Deploying HTTPS allows the use of HTTP/2, that are new generations of HTTP, designed to reduce page load times and latency. It is recommended to use HTTP Strict Transport Security with HTTPS to protect users from man-in-the-middle attacks SSL stripping. HTTPS should not be confused with the little-used Secure HTTP specified in RFC 2660; as of April 2018, 33.2% of Alexa top 1,000,000 websites use HTTPS as default, 57.1% of the Internet's 137,971 most popular websites have a secure implementation of HTTPS, 70% of page loads use HTTPS. Most browsers display a warning. Older browsers, when connecting to a site wit
Home theater PC
A home theater PC or media center computer is a convergence device that combines some or all the capabilities of a personal computer with a software application that supports video, audio playback, sometimes video recording functionality. In recent years, other types of consumer electronics, including gaming systems and dedicated media devices have crossed over to manage video and music content; the term "media center" refers to specialized application software designed to run on standard personal computers. An HTPC and other convergence devices integrate components of a home theater into a unit co-located with a home entertainment system. An HTPC system has a remote control and the software interface has a 10-foot user interface design so that it can be comfortably viewed at typical television viewing distances. An HTPC can be purchased pre-configured with the required hardware and software needed to add video programming or music to the PC. Enthusiasts can piece together a system out of discrete components as part of a software-based HTPC.
Since 2007 Digital media player and Smart TV software has been incorporated into consumer electronics through software or hardware changes including video game consoles, Blu-ray players, networked media players and set-top boxes. The increased availability of specialized devices, coupled with paid and free digital online content, now offer alternatives to multipurpose personal computers; the HTPC as a concept is the product of several technology innovations including high-powered home computers, digital media, the shift from standard resolution CRT to high definition monitors and large screen televisions. Integrating televisions and personal computers dates back to the late 1980s with tuner cards that could be added to Commodore Amiga PCs via the Video Toaster; this adaptation would allow a small video window to appear on the screen with broadcast or cable content. Apple Computer developed the Macintosh TV in late 1993 that included a tuner card built into a Macintosh LC 520 chassis but withdrew from the market with only 10,000 units shipped.
In 1996 Gateway Computer unveiled the Destination computer that included a tuner card and video card. The unit cost $4,000 and integrated television viewing and computer functions on one color monitor; the Destination was called a "PC-TV Combo" but by December the term "Home-theater PC" appeared in mainstream media: "The home theater PC will be a combination entertainment and information appliance". By 2000, DVD players had become ubiquitous and consumers were seeking ways to improve the picture; the value of using a computer instead of stand alone DVD player drove more usage of the PC as a home media device. In particular, the desire for progressive scanning DVD players with better video fidelity led some consumers to consider their computers instead of expensive DVD players; as DVD players dropped in price, so did PCs and their related video processing and storage capabilities. In 2000, DVD decryption software using the DeCSS algorithm let DVD owners consolidate their DVD video libraries on hard-drives.
Innovations like TiVo and ReplayTV allowed viewers to store and timeshift broadcast content using specialty designed computers. ReplayTV for instance ran on a VxWorks platform. Incorporating these capabilities into PCs was well within the ability of a computer hobbyist, willing to build and program these systems. Key benefits of these DIY projects included more features. Advancements in hardware identified another weak link: the absence of media management software to make it easy to display and control the video from a distance. By 2002, major software developments facilitated media management, hardware integration, content presentation. MythTV provided a open source solution using Linux; the concept was to combine a digital tuner with digital video recording, program guides, computer capabilities with a 10-foot user interface. XBMC was another free and open software project started with re-purposing the Xbox as a home theater PC but has since been ported to Windows and Macintosh operating systems in various forms including Boxee and Plex.
Mainstream commercial software packages included Microsoft's XP Media Center Edition, bundled with Windows XP and Apple's Front Row software bundled with Mac OS X. By early 2006, commercial examples of this integration included the Mac mini which had the Apple Remote, 5.1 digital audio, an updated Front Row interface that would play shared media. Because of these features and the Mini's small form factor, consumers began using the Mini as a Mac-based home theater PC; as digital cable and satellite became the norm, HTPC software became more dependent on external decoder boxes, the subscription costs that came with them. For instance, MythTV is capable of capturing unencrypted HDTV streams, such as those broadcast over the air or on cable using a QAM tuner. However, most U. S. cable and satellite set-top boxes provide only encrypted HD streams for "non-basic" content, which can be decoded only by OpenCable-approved hardware or software. In September 2009, OEM restrictions were lifted for cableCARD devices, opening up the possibility for HTPC integration.
The advent of the digital HDTV displays helped to complete the value and ease of use of a HTPC system. Digital projectors, plasma and LCD displays came pre-configured to accept computer video outputs including VGA, DVI and Component Video. Furthermore, both the computers and the displays could include video scalers to better conform the image to the screen format and resolutions. Computers included HDMI ports that carry both audio and video sig
Hypertext Transfer Protocol
The Hypertext Transfer Protocol is an application protocol for distributed, hypermedia information systems. HTTP is the foundation of data communication for the World Wide Web, where hypertext documents include hyperlinks to other resources that the user can access, for example by a mouse click or by tapping the screen in a web browser. HTTP was developed to facilitate the World Wide Web. Development of HTTP was initiated by Tim Berners-Lee at CERN in 1989. Development of HTTP standards was coordinated by the Internet Engineering Task Force and the World Wide Web Consortium, culminating in the publication of a series of Requests for Comments; the first definition of HTTP/1.1, the version of HTTP in common use, occurred in RFC 2068 in 1997, although this was made obsolete by RFC 2616 in 1999 and again by the RFC 7230 family of RFCs in 2014. A version, the successor HTTP/2, was standardized in 2015, is now supported by major web servers and browsers over Transport Layer Security using Application-Layer Protocol Negotiation extension where TLS 1.2 or newer is required.
HTTP functions as a request–response protocol in the client–server computing model. A web browser, for example, may be the client and an application running on a computer hosting a website may be the server; the client submits an HTTP request message to the server. The server, which provides resources such as HTML files and other content, or performs other functions on behalf of the client, returns a response message to the client; the response contains completion status information about the request and may contain requested content in its message body. A web browser is an example of a user agent. Other types of user agent include the indexing software used by search providers, voice browsers, mobile apps, other software that accesses, consumes, or displays web content. HTTP is designed to permit intermediate network elements to improve or enable communications between clients and servers. High-traffic websites benefit from web cache servers that deliver content on behalf of upstream servers to improve response time.
Web browsers cache accessed web resources and reuse them, when possible, to reduce network traffic. HTTP proxy servers at private network boundaries can facilitate communication for clients without a globally routable address, by relaying messages with external servers. HTTP is an application layer protocol designed within the framework of the Internet protocol suite, its definition presumes an underlying and reliable transport layer protocol, Transmission Control Protocol is used. However, HTTP can be adapted to use unreliable protocols such as the User Datagram Protocol, for example in HTTPU and Simple Service Discovery Protocol. HTTP resources are identified and located on the network by Uniform Resource Locators, using the Uniform Resource Identifiers schemes http and https. URIs and hyperlinks in HTML documents form interlinked hypertext documents. HTTP/1.1 is a revision of the original HTTP. In HTTP/1.0 a separate connection to the same server is made for every resource request. HTTP/1.1 can reuse a connection multiple times to download images, stylesheets, etc after the page has been delivered.
HTTP/1.1 communications therefore experience less latency as the establishment of TCP connections presents considerable overhead. The term hypertext was coined by Ted Nelson in 1965 in the Xanadu Project, in turn inspired by Vannevar Bush's 1930s vision of the microfilm-based information retrieval and management "memex" system described in his 1945 essay "As We May Think". Tim Berners-Lee and his team at CERN are credited with inventing the original HTTP, along with HTML and the associated technology for a web server and a text-based web browser. Berners-Lee first proposed the "WorldWideWeb" project in 1989—now known as the World Wide Web; the first version of the protocol had only one method, namely GET, which would request a page from a server. The response from the server was always an HTML page; the first documented version of HTTP was HTTP V0.9. Dave Raggett led the HTTP Working Group in 1995 and wanted to expand the protocol with extended operations, extended negotiation, richer meta-information, tied with a security protocol which became more efficient by adding additional methods and header fields.
RFC 1945 introduced and recognized HTTP V1.0 in 1996. The HTTP WG planned to publish new standards in December 1995 and the support for pre-standard HTTP/1.1 based on the developing RFC 2068 was adopted by the major browser developers in early 1996. By March that year, pre-standard HTTP/1.1 was supported in Arena, Netscape 2.0, Netscape Navigator Gold 2.01, Mosaic 2.7, Lynx 2.5, in Internet Explorer 2.0. End-user adoption of the new browsers was rapid. In March 1996, one web hosting company reported that over 40% of browsers in use on the Internet were HTTP 1.1 compliant. That same web hosting company reported that by June 1996, 65% of all browsers accessing their servers were HTTP/1.1 compliant. The HTTP/1.1 standard as defined in RFC 2068 was released in January 1997. Improvements and updates to the HTTP/1.1 standard were released under RFC 2616 in June 1999. In 2007, the HTTPbis Working Group was formed, in part, to revise and clarify the HTTP/1.1 specification. In June 2014, the WG released an updated six-part specification obsoleting RFC 2616: RFC 7230, HTTP/1.1: Message Syntax and Routing RFC 7231, HTTP/1.1: Semantics and Content RFC 7232, HTTP/1.1: Conditional Requests RFC 7233, HTTP/1.1: Range Requests RFC 7234, HTTP/1.1: Caching RFC 7235, HTTP/1
Windows XP is a personal computer operating system produced by Microsoft as part of the Windows NT family of operating systems. It was released to manufacturing on August 24, 2001, broadly released for retail sale on October 25, 2001. Development of Windows XP began in the late 1990s as "Neptune", an operating system built on the Windows NT kernel, intended for mainstream consumer use. An updated version of Windows 2000 was originally planned for the business market; as such, Windows XP was the first consumer edition of Windows not to be based on MS-DOS. Upon its release, Windows XP received positive reviews, with critics noting increased performance and stability, a more intuitive user interface, improved hardware support, expanded multimedia capabilities. However, some industry reviewers were concerned by the new licensing model and product activation system. Extended support for Windows XP ended on April 8, 2014, after which the operating system ceased receiving further support or security updates to most users.
As of March 2019, 1.75% of Windows PCs run Windows XP, the OS is still most popular in some countries with up to 38% of the Windows share. In the late 1990s, initial development of what would become Windows XP was focused on two individual products. However, the projects proved to be too ambitious. In January 2000, shortly prior to the official release of Windows 2000, technology writer Paul Thurrott reported that Microsoft had shelved both Neptune and Odyssey in favor of a new product codenamed "Whistler", after Whistler, British Columbia, as many Microsoft employees skied at the Whistler-Blackcomb ski resort; the goal of Whistler was to unify both the consumer and business-oriented Windows lines under a single, Windows NT platform: Thurrott stated that Neptune had become "a black hole when all the features that were cut from were re-tagged as Neptune features. And since Neptune and Odyssey would be based on the same code-base anyway, it made sense to combine them into a single project". At PDC on July 13, 2000, Microsoft announced that Whistler would be released during the second half of 2001, unveiled the first preview build, 2250.
The build notably introduced an early version of Windows XP's visual styles system. Microsoft released the first beta build of Whistler, build 2296, on October 31, 2000. Subsequent builds introduced features that users of the release version of Windows XP would recognise, such as Internet Explorer 6.0, the Microsoft Product Activation system and the Bliss desktop background. On February 5, 2001, Microsoft announced that Whistler would be known as Windows XP, where XP stands for "eXPerience". In June 2001, Microsoft indicated that it was planning to, in conjunction with Intel and other PC makers, spend at least 1 billion US dollars on marketing and promoting Windows XP; the theme of the campaign, "Yes You Can", was designed to emphasize the platform's overall capabilities. Microsoft had planned to use the slogan "Prepare to Fly", but it was replaced due to sensitivity issues in the wake of the September 11 attacks. On August 24, 2001, Windows XP build. During a ceremonial media event at Microsoft Redmond Campus, copies of the RTM build were given to representatives of several major PC manufacturers in briefcases, who flew off on decorated helicopters.
While PC manufacturers would be able to release devices running XP beginning on September 24, 2001, XP was expected to reach general, retail availability on October 25, 2001. On the same day, Microsoft announced the final retail pricing of XP's two main editions, "Home" and "Professional". While retaining some similarities to previous versions, Windows XP's interface was overhauled with a new visual appearance, with an increased use of alpha compositing effects, drop shadows, "visual styles", which changed the appearance of the operating system; the number of effects enabled are determined by the operating system based on the computer's processing power, can be enabled or disabled on a case-by-case basis. XP added ClearType, a new subpixel rendering system designed to improve the appearance of fonts on liquid-crystal displays. A new set of system icons was introduced; the default wallpaper, Bliss, is a photo of a landscape in the Napa Valley outside Napa, with rolling green hills and a blue sky with stratocumulus and cirrus clouds.
The Start menu received its first major overhaul in XP, switching to a two-column layout with the ability to list and display used applications opened documents, the traditional cascading "All Programs" menu. The taskbar can now group windows opened by a single application into one taskbar button, with a popup menu listing the individual windows; the notification area hides "inactive" icons by default. A "common tasks" list was added, Windows Explorer's sidebar was updated to use a new task-based design with lists of common actions. Fast user switching allows additional users to log into a Windows XP machine without existing users having to close their programs and loggin