Australian National University
The Australian National University is a national research university located in Canberra, the capital of Australia. Its main campus in Acton encompasses seven teaching and research colleges, in addition to several national academies and institutes. Founded in 1946, it is the only university to have been created by the Parliament of Australia. A postgraduate research university, ANU commenced undergraduate teaching in 1960 when it integrated the Canberra University College, established in 1929 as a campus of the University of Melbourne. ANU employs 3,753 staff; the university's endowment stood at A$1.13 billion in 2012. ANU is regarded as one of the world's leading research universities, it is ranked 1st in Australia and the whole of Oceania, 24th in the world by the 2019 QS World University Rankings, 49th in the world by the 2019 Times Higher Education. ANU was named the world's 7th most international university in a 2017 study by Times Higher Education. In the 2017 Times Higher Education Global Employability University Ranking, an annual ranking of university graduates' employability, ANU was ranked 21st in the world.
ANU is ranked 100th in the CWTS Leiden ranking. The university is well known for its programmes in the arts and social sciences, ranks among the best in the world for a number of disciplines including politics and international relations, social policy, geography. ANU counts six Nobel laureates and 49 Rhodes scholars among its faculty and alumni; the university has educated two prime ministers, 30 current Australian ambassadors and more than a dozen current heads of government departments of Australia. The latest releases of ANU's scholarly publications are held through ANU Press online. Calls for the establishment of a national university in Australia began as early as 1900. After the location of the nation's capital, was determined in 1908, land was set aside for the university at the foot of Black Mountain in the city designs by Walter Burley Griffin. Planning for the university was disrupted by World War II but resumed with the creation of the Department of Post-War Reconstruction in 1942 leading to the passage of the Australian National University Act 1946 by the Chifley Government on 1 August 1946.
A group of eminent Australian scholars returned from overseas to join the university, including Sir Howard Florey, Sir Mark Oliphant, Sir Keith Hancock and Sir Raymond Firth. Economist Sir Douglas Copland was appointed as ANU's first Vice-Chancellor and former Prime Minister Stanley Bruce served as the first Chancellor. ANU was organised into four centres—the Research Schools of Physical Sciences, Social Sciences and Pacific Studies and the John Curtin School of Medical Research; the first residents' hall, University House, was opened in 1954 for faculty members and postgraduate students. Mount Stromlo Observatory, established by the federal government in 1924, became part of ANU in 1957; the first locations of the ANU Library, the Menzies and Chifley buildings, opened in 1963. The Australian Forestry School, located in Canberra since 1927, was amalgamated by ANU in 1965. Canberra University College was the first institution of higher education in the national capital, having been established in 1929 and enrolling its first undergraduate pupils in 1930.
Its founding was led by Sir Robert Garran, one of the drafters of the Australian Constitution and the first Solicitor-General of Australia. CUC was affiliated with the University of Melbourne and its degrees were granted by that university. Academic leaders at CUC included historian Manning Clark, political scientist Finlay Crisp, poet A. D. Hope and economist Heinz Arndt. In 1960, CUC was integrated into ANU as the School of General Studies with faculties in arts, economics and science. Faculties in Oriental studies and engineering were introduced later. Bruce Hall, the first residential college for undergraduates, opened in 1961; the Canberra School of Music and the Canberra School of Art combined in 1988 to form the Canberra Institute of the Arts, amalgamated with the university as the ANU Institute of the Arts in 1992. ANU established its Medical School in 2002, after obtaining federal government approval in 2000. On 18 January 2003, the Canberra bushfires destroyed the Mount Stromlo Observatory.
ANU astronomers now conduct research from the Siding Spring Observatory, which contains 10 telescopes including the Anglo-Australian Telescope. In February 2013, financial entrepreneur and ANU graduate Graham Tuckwell made the largest university donation in Australian history by giving $50 million to fund an undergraduate scholarship program at ANU. ANU is well known for its history of student activism and, in recent years, its fossil fuel divestment campaign, one of the longest-running and most successful in the country; the decision of the ANU Council to divest from two fossil fuel companies in 2014 was criticised by ministers in the Abbott government, but defended by Vice Chancellor Ian Young, who noted:On divestment, it is clear we were in the right and played a national and international leadership role. E seem to have played a major role in a movement; as of 2014 ANU still had investments in major fossil fuel companies. A survey conducted by the Australian Human Rights Commission in 2017 found that the ANU had the second highest incidence of sexual assault and sexual harassment.
3.5 per cent of respondents from the ANU re
Internet protocol suite
The Internet protocol suite is the conceptual model and set of communications protocols used in the Internet and similar computer networks. It is known as TCP/IP because the foundational protocols in the suite are the Transmission Control Protocol and the Internet Protocol, it is known as the Department of Defense model because the development of the networking method was funded by the United States Department of Defense through DARPA. The Internet protocol suite provides end-to-end data communication specifying how data should be packetized, transmitted and received; this functionality is organized into four abstraction layers, which classify all related protocols according to the scope of networking involved. From lowest to highest, the layers are the link layer, containing communication methods for data that remains within a single network segment; the technical standards underlying the Internet protocol suite and its constituent protocols are maintained by the Internet Engineering Task Force.
The Internet protocol suite predates the OSI model, a more comprehensive reference framework for general networking systems. The Internet protocol suite resulted from research and development conducted by the Defense Advanced Research Projects Agency in the late 1960s. After initiating the pioneering ARPANET in 1969, DARPA started work on a number of other data transmission technologies. In 1972, Robert E. Kahn joined the DARPA Information Processing Technology Office, where he worked on both satellite packet networks and ground-based radio packet networks, recognized the value of being able to communicate across both. In the spring of 1973, Vinton Cerf, who helped develop the existing ARPANET Network Control Program protocol, joined Kahn to work on open-architecture interconnection models with the goal of designing the next protocol generation for the ARPANET. By the summer of 1973, Kahn and Cerf had worked out a fundamental reformulation, in which the differences between local network protocols were hidden by using a common internetwork protocol, instead of the network being responsible for reliability, as in the ARPANET, this function was delegated to the hosts.
Cerf credits Hubert Zimmermann and Louis Pouzin, designer of the CYCLADES network, with important influences on this design. The protocol was implemented as the Transmission Control Program, first published in 1974; the TCP managed both datagram transmissions and routing, but as the protocol grew, other researchers recommended a division of functionality into protocol layers. Advocates included Jonathan Postel of the University of Southern California's Information Sciences Institute, who edited the Request for Comments, the technical and strategic document series that has both documented and catalyzed Internet development. Postel stated, "We are screwing up in our design of Internet protocols by violating the principle of layering." Encapsulation of different mechanisms was intended to create an environment where the upper layers could access only what was needed from the lower layers. A monolithic design would lead to scalability issues; the Transmission Control Program was split into two distinct protocols, the Transmission Control Protocol and the Internet Protocol.
The design of the network included the recognition that it should provide only the functions of efficiently transmitting and routing traffic between end nodes and that all other intelligence should be located at the edge of the network, in the end nodes. This design is known as the end-to-end principle. Using this design, it became possible to connect any network to the ARPANET, irrespective of the local characteristics, thereby solving Kahn's initial internetworking problem. One popular expression is that TCP/IP, the eventual product of Cerf and Kahn's work, can run over "two tin cans and a string." Years as a joke, the IP over Avian Carriers formal protocol specification was created and tested. A computer called, it forwards network packets forth between them. A router was called gateway, but the term was changed to avoid confusion with other types of gateways. From 1973 to 1974, Cerf's networking research group at Stanford worked out details of the idea, resulting in the first TCP specification.
A significant technical influence was the early networking work at Xerox PARC, which produced the PARC Universal Packet protocol suite, much of which existed around that time. DARPA contracted with BBN Technologies, Stanford University, the University College London to develop operational versions of the protocol on different hardware platforms. Four versions were developed: TCP v1, TCP v2, TCP v3 and IP v3, TCP/IP v4; the last protocol is still in use today. In 1975, a two-network TCP/IP communications test was performed between Stanford and University College London. In November 1977, a three-network TCP/IP test was conducted between sites in the US, the UK, Norway. Several other TCP/IP prototypes were developed at multiple research centers between 1978 and 1983. In March 1982, the US Department of Defense declared TCP/IP as the standard for all military computer networking. In the same year, Peter T. Kirstein's research group at University College London adopted the protocol; the migration of the ARPANET to TCP/IP was completed on flag day January 1, 1983, when the new protocols were permanently activated.
In 1985, the Internet Advisory Board held a three-day TCP/
A computer file is a computer resource for recording data discretely in a computer storage device. Just as words can be written to paper, so can information be written to a computer file. Files can be transferred through the internet. There are different types of computer files, designed for different purposes. A file may be designed to store a picture, a written message, a video, a computer program, or a wide variety of other kinds of data; some types of files can store several types of information at once. By using computer programs, a person can open, change and close a computer file. Computer files may be reopened and copied an arbitrary number of times. Files are organised in a file system, which keeps track of where the files are located on disk and enables user access; the word "file" derives from the Latin filum."File" was used in the context of computer storage as early as January 1940. In Punched Card Methods in Scientific Computation, W. J. Eckert stated, "The first extensive use of the early Hollerith Tabulator in astronomy was made by Comrie.
He used it for building a table from successive differences, for adding large numbers of harmonic terms". "Tables of functions are constructed from their differences with great efficiency, either as printed tables or as a file of punched cards." In February 1950: In a Radio Corporation of America advertisement in Popular Science Magazine describing a new "memory" vacuum tube it had developed, RCA stated: "the results of countless computations can be kept'on file' and taken out again. Such a'file' now exists in a'memory' tube developed at RCA Laboratories. Electronically it retains figures fed into calculating machines, holds them in storage while it memorizes new ones - speeds intelligent solutions through mazes of mathematics." In 1952, "file" denoted, information stored on punched cards. In early use, the underlying hardware, rather than the contents stored on it, was denominated a "file". For example, the IBM 350 disk drives were denominated "disk files"; the introduction, circa 1961, by the Burroughs MCP and the MIT Compatible Time-Sharing System of the concept of a "file system" that managed several virtual "files" on one storage device is the origin of the contemporary denotation of the word.
Although the contemporary "register file" demonstrates the early concept of files, its use has decreased. On most modern operating systems, files are organized into one-dimensional arrays of bytes; the format of a file is defined by its content since a file is a container for data, although, on some platforms the format is indicated by its filename extension, specifying the rules for how the bytes must be organized and interpreted meaningfully. For example, the bytes of a plain text file are associated with either ASCII or UTF-8 characters, while the bytes of image and audio files are interpreted otherwise. Most file types allocate a few bytes for metadata, which allows a file to carry some basic information about itself; some file systems can store arbitrary file-specific data outside of the file format, but linked to the file, for example extended attributes or forks. On other file systems this can be done via software-specific databases. All those methods, are more susceptible to loss of metadata than are container and archive file formats.
At any instant in time, a file might have a size expressed as number of bytes, that indicates how much storage is associated with the file. In most modern operating systems the size can be any non-negative whole number of bytes up to a system limit. Many older operating systems kept track only of the number of blocks or tracks occupied by a file on a physical storage device. In such systems, software employed other methods to track the exact byte count; the general definition of a file does not require that its size have any real meaning, unless the data within the file happens to correspond to data within a pool of persistent storage. A special case is a zero byte file. For example, the file to which the link /bin/ls points in a typical Unix-like system has a defined size that changes. Compare this with /dev/null, a file, but its size may be obscure. Information in a computer file can consist of smaller packets of information that are individually different but share some common traits. For example, a payroll file might contain information concerning all the employees in a company and their payroll details.
A text file may contain lines of corresponding to printed lines on a piece of paper. Alternatively, a file may contain an arbitrary binary image or it may contain an executable; the way information is grouped into a file is up to how it is designed. This has led to a plethora of more or less standardized file structures for all imaginable purposes, from the simplest to the most complex. Most computer files are used by computer programs which create, modify or delete the files for their own use on an as-needed basis; the programmers who create the programs decide what files are needed, how they are to be used and their names. In some cases, computer pr
Andrew "Tridge" Tridgell is an Australian computer programmer. He is the author of and a contributor to the Samba file server, co-inventor of the rsync algorithm, he has analysed complex proprietary protocols and algorithms, to allow compatible free and open source software implementations. Tridgell was a major developer of the Samba software, analyzing the Server Message Block protocol used for workgroup and network file sharing by Microsoft Windows products, he developed the talloc hierarchical memory allocator as part of Samba. For his PhD dissertation, he co-developed rsync, including the rsync algorithm, a efficient file transfer and synchronisation tool, he was the original author of rzip, which uses a similar algorithm to rsync. He developed spamsum, based on locality-sensitive hashing algorithms, he is the author of a reinforcement-learning based chess engine. Tridgell was a leader in hacking the TiVo to make it work in Australia, which uses the PAL video format. In April 2005, Tridgell tried to produce free software that interoperated with the BitKeeper source code repository.
This was cited as the reason that BitMover revoked a license allowing Linux developers free use of their BitKeeper product. Linus Torvalds, the creator of the Linux kernel, Tridgell were thus involved in a public debate about the events, in which Tridgell stated that, not having bought or owned BitKeeper – and thus having never agreed to its license – he could not violate it, was analyzing the protocol ethically, as he had done with Samba. Tridgell's involvement in the project resulted in Torvalds accusing him of playing dirty tricks with BitKeeper. Tridgell claimed his analysis started with telneting to a BitKeeper server and typing help. In 2011 Tridgell got involved with the software development of ArduPilot Mega, an open source Arduino-based UAV controller board, working on an entry for the UAV Challenge Outback Rescue. Tridgell completed a PhD at the Computer Sciences Laboratory of the Australian National University, his original doctorate work was never completed. His submitted dissertation'Efficient Algorithms for Sorting and Synchronization' was based on his work on the rsync algorithm.
In October 2003, The Bulletin magazine judged Tridgell to be Australia's smartest Information and Communications Technology person. In July 2008, Tridgell was named "Best Interoperator" at the Google-O'Reilly Open Source Awards, for his work on Samba and Rsync. Tridgell has been called a "guru in its traditional Indian meaning" by Sam Varghese. On 11 December 2018, Tridgell was awarded the degree of Doctor of Science by the Australian National University, for authoring Samba, co-inventing rsync. Andrew Tridgell's website Efficient Algorithms for Sorting and Synchronization – Active Directory in Samba 4'an old story' FOSS folk who make us proud Patent Defence for Free Software, January 2010 presentation transcript
Jeremy Allison is a computer programmer known for his contributions to the free software community, notably to Samba, a re-implementation of SMB/CIFS networking protocol, released under the GNU General Public License. Other contributions include the early versions of the pwdump password cracking utility. During his career, Jeremy Allison has defended the free software approach: He pitched making Vantive code free software to its founder, he persuaded Michael Tiemann to use the GNU General Public License for Cygwin. He convinced Tim Wilkinson to put the Kaffe virtual machine for Java under the GPL, he was involved in Silicon Graphics' decision to put XFS for Linux under the GPL. This commitment to free software culminated with his decision to leave Novell in protest of a patent deal, considered by many as a FUD attack on Linux and other free software, by Allison as breaking section 7 of the GNU General Public License. Jeremy Allison's web page
Berkeley Software Distribution
The Berkeley Software Distribution was an operating system based on Research Unix and distributed by the Computer Systems Research Group at the University of California, Berkeley. Today, "BSD" refers to its descendants, such as FreeBSD, OpenBSD, NetBSD, or DragonFly BSD. BSD was called Berkeley Unix because it was based on the source code of the original Unix developed at Bell Labs. In the 1980s, BSD was adopted by workstation vendors in the form of proprietary Unix variants such as DEC Ultrix and Sun Microsystems SunOS due to its permissive licensing and familiarity to many technology company founders and engineers. Although these proprietary BSD derivatives were superseded in the 1990s by UNIX SVR4 and OSF/1 releases provided the basis for several open-source operating systems including FreeBSD, OpenBSD, NetBSD, DragonFly BSD, TrueOS. These, in turn, have been used by proprietary operating systems, including Apple's macOS and iOS, which derived from them, Microsoft Windows, which used a part of its TCP/IP code.
The earliest distributions of Unix from Bell Labs in the 1970s included the source code to the operating system, allowing researchers at universities to modify and extend Unix. The operating system arrived at Berkeley in 1974, at the request of computer science professor Bob Fabry, on the program committee for the Symposium on Operating Systems Principles where Unix was first presented. A PDP-11/45 was bought to run the system, but for budgetary reasons, this machine was shared with the mathematics and statistics groups at Berkeley, who used RSTS, so that Unix only ran on the machine eight hours per day. A larger PDP-11/70 was installed at Berkeley the following year, using money from the Ingres database project. In 1975, Ken Thompson came to Berkeley as a visiting professor, he started working on a Pascal implementation for the system. Graduate students Chuck Haley and Bill Joy improved Thompson's Pascal and implemented an improved text editor, ex. Other universities became interested in the software at Berkeley, so in 1977 Joy started compiling the first Berkeley Software Distribution, released on March 9, 1978.
1BSD was an add-on to Version 6 Unix rather than a complete operating system in its own right. Some thirty copies were sent out; the second Berkeley Software Distribution, released in May 1979, included updated versions of the 1BSD software as well as two new programs by Joy that persist on Unix systems to this day: the vi text editor and the C shell. Some 75 copies of 2BSD were sent out by Bill Joy. A VAX computer was installed at Berkeley in 1978, but the port of Unix to the VAX architecture, UNIX/32V, did not take advantage of the VAX's virtual memory capabilities; the kernel of 32V was rewritten by Berkeley students to include a virtual memory implementation, a complete operating system including the new kernel, ports of the 2BSD utilities to the VAX, the utilities from 32V was released as 3BSD at the end of 1979. 3BSD was alternatively called Virtual VAX/UNIX or VMUNIX, BSD kernel images were called /vmunix until 4.4BSD. After 4.3BSD was released in June 1986, it was determined that BSD would move away from the aging VAX platform.
The Power 6/32 platform developed by Computer Consoles Inc. seemed promising at the time, but was abandoned by its developers shortly thereafter. Nonetheless, the 4.3BSD-Tahoe port proved valuable, as it led to a separation of machine-dependent and machine-independent code in BSD which would improve the system's future portability. In addition to portability, the CSRG worked on an implementation of the OSI network protocol stack, improvements to the kernel virtual memory system and new TCP/IP algorithms to accommodate the growth of the Internet; until all versions of BSD used proprietary AT&T Unix code, were therefore subject to an AT&T software license. Source code licenses had become expensive and several outside parties had expressed interest in a separate release of the networking code, developed outside AT&T and would not be subject to the licensing requirement; this led to Networking Release 1, made available to non-licensees of AT&T code and was redistributable under the terms of the BSD license.
It was released in June 1989. After Net/1, BSD developer Keith Bostic proposed that more non-AT&T sections of the BSD system be released under the same license as Net/1. To this end, he started a project to reimplement most of the standard Unix utilities without using the AT&T code. Within eighteen months, all of the AT&T utilities had been replaced, it was determined that only a few AT&T files remained in the kernel; these files were removed, the result was the June 1991 release of Networking Release 2, a nearly complete operating system, distributable. Net/2 was the basis for two separate ports of BSD to the Intel 80386 architecture: the free 386BSD by William Jolitz and the proprietary BSD/386 by Berkeley Software Design. 386BSD itself was short-lived, but became the initial code base of the NetBSD and FreeBSD projects that were started shortly thereafter. BSDi soon found itself in legal trouble with AT&T's Unix System Laboratories subsidiary the owners of the System V copyright and the Unix trademark.
The USL v. BSDi lawsuit was filed in 1992 and led to an injunction on the distribution of Net/2 until the validity of USL's copyright claims on the source could be determined; the lawsuit slowed development of the free-
An operating system is system software that manages computer hardware and software resources and provides common services for computer programs. Time-sharing operating systems schedule tasks for efficient use of the system and may include accounting software for cost allocation of processor time, mass storage and other resources. For hardware functions such as input and output and memory allocation, the operating system acts as an intermediary between programs and the computer hardware, although the application code is executed directly by the hardware and makes system calls to an OS function or is interrupted by it. Operating systems are found on many devices that contain a computer – from cellular phones and video game consoles to web servers and supercomputers; the dominant desktop operating system is Microsoft Windows with a market share of around 82.74%. MacOS by Apple Inc. is in second place, the varieties of Linux are collectively in third place. In the mobile sector, use in 2017 is up to 70% of Google's Android and according to third quarter 2016 data, Android on smartphones is dominant with 87.5 percent and a growth rate 10.3 percent per year, followed by Apple's iOS with 12.1 percent and a per year decrease in market share of 5.2 percent, while other operating systems amount to just 0.3 percent.
Linux distributions are dominant in supercomputing sectors. Other specialized classes of operating systems, such as embedded and real-time systems, exist for many applications. A single-tasking system can only run one program at a time, while a multi-tasking operating system allows more than one program to be running in concurrency; this is achieved by time-sharing, where the available processor time is divided between multiple processes. These processes are each interrupted in time slices by a task-scheduling subsystem of the operating system. Multi-tasking may be characterized in co-operative types. In preemptive multitasking, the operating system slices the CPU time and dedicates a slot to each of the programs. Unix-like operating systems, such as Solaris and Linux—as well as non-Unix-like, such as AmigaOS—support preemptive multitasking. Cooperative multitasking is achieved by relying on each process to provide time to the other processes in a defined manner. 16-bit versions of Microsoft Windows used cooperative multi-tasking.
32-bit versions of both Windows NT and Win9x, used preemptive multi-tasking. Single-user operating systems have no facilities to distinguish users, but may allow multiple programs to run in tandem. A multi-user operating system extends the basic concept of multi-tasking with facilities that identify processes and resources, such as disk space, belonging to multiple users, the system permits multiple users to interact with the system at the same time. Time-sharing operating systems schedule tasks for efficient use of the system and may include accounting software for cost allocation of processor time, mass storage and other resources to multiple users. A distributed operating system manages a group of distinct computers and makes them appear to be a single computer; the development of networked computers that could be linked and communicate with each other gave rise to distributed computing. Distributed computations are carried out on more than one machine; when computers in a group work in cooperation, they form a distributed system.
In an OS, distributed and cloud computing context, templating refers to creating a single virtual machine image as a guest operating system saving it as a tool for multiple running virtual machines. The technique is used both in virtualization and cloud computing management, is common in large server warehouses. Embedded operating systems are designed to be used in embedded computer systems, they are designed to operate on small machines like PDAs with less autonomy. They are able to operate with a limited number of resources, they are compact and efficient by design. Windows CE and Minix 3 are some examples of embedded operating systems. A real-time operating system is an operating system that guarantees to process events or data by a specific moment in time. A real-time operating system may be single- or multi-tasking, but when multitasking, it uses specialized scheduling algorithms so that a deterministic nature of behavior is achieved. An event-driven system switches between tasks based on their priorities or external events while time-sharing operating systems switch tasks based on clock interrupts.
A library operating system is one in which the services that a typical operating system provides, such as networking, are provided in the form of libraries and composed with the application and configuration code to construct a unikernel: a specialized, single address space, machine image that can be deployed to cloud or embedded environments. Early computers were built to perform a series of single tasks, like a calculator. Basic operating system features were developed in the 1950s, such as resident monitor functions that could automatically run different programs in succession to speed up processing. Operating systems did not exist in their more complex forms until the early 1960s. Hardware features were added, that enabled use of runtime libraries and parallel processing; when personal computers became popular in the 1980s, operating systems were made for them similar in concept to those used on larger computers. In the 1940s, the earliest electronic digital systems had no operating systems.
Electronic systems of this time were programmed on rows of mechanical switches or by jumper wires on plug boards. These were special-purpose systems that, for example, generated ballistics tables for the military or controlled the pri