Introduction
Linux OS was first created by a student from the University of Helsinki in Finland. The creator’s name was Linus Torvalds and he had an interest which turned into a passion for Minix, a small Unix application which was later developed into a system that surpassed the Minix standards. He started working on the minix in 1991 and worked heavily until 1994 when the first version of Linux kernal 1.0 was released. This Linux kernal sets the foundation to which the OS of Linux is formed. Hundreds of organizations and companies today have hired individuals and used them to release versions of operating systems using Linux kernal.
Linux’s functioning, features and adaptation have made Linux and Windows OS’s are excellent alternatives to other OS’s. IBM and other giant companies around the world support Linux and its ongoing work after a decade from its initial release. The OS is incorporated into microchips using a process called “embedding” and is increasing the performance of of appliances and devices.
History of Linux
Through the 1990’s some computer savy technicians and hobby insistent people with an interest in computers developed desktop management systems. These systems including GNOME and KDE that run on applications on Linux are available to anyone regardless of the persons motive to use the system. Linus Torvalds was interested in learning the capabilities and features of an 80386 processor for task switching. The application originally named Freax was first used with the Minix operating system.
Both the Freax and Minix designs seemed to be sacrificing performance for academic research and studying. Many of the computing specialists now are making assumptions that have changed since the 90’s. Portability is now a common goal for these specialists of the computer industry and this is certainly not a academic requirement for software. Various ports to IA-32, PowerPC, MIPS, Alpha, and ARM along with supporting products being made and sold to wholesalers and retailers, commercial enterprises gave Linus a Alpha based system when tasks on Linus’s priority list moved up to a notably busy point.
History of Windows
Presidents of Microsoft were Bill Gates and Paul Allen they shared the title until 1977, when Bill Gates became president and Paul Allen vice president. In 1978 the disk drives of the Tandy and Apple machines were 5.25-inch. First COMDEX computer show in Las Vegas introduces a 16-bit microprocessor, and from Intel manufacturers they introduce a 8086 chip. Al Gore comes up with the phrase “information highway.” The same year Apple co-founder Steve Wozniak developed the first programming language called Integer Basic, this language was quickly replaced by the Microsoft Applesoft Basic.
Also in 1978, there was a machine that had an integrated, self contained design and was priced at less than $800, known as the Commodore PET which was a Personal Electronic Transactor. On 4/11/78 Microsoft announces its third language product, Microsoft COBOL-80. On the 1st of November in 1978 after their third language introduction, they opened their first international sales office in Japan. Microsoft delegates ASCII Microsoft, locatwed in Tokyo, asits exclusive sales agent for the Far East. And finally on New Years Eve of 1978 Microsoft announced that their year end sales was over $1 million dollars. The following year in April of 1979 Microsoft 8080 BASIC is the first microprocessor to win the ICP Million Dollar Award. The big computers were dominated by software for the mainframe computer, the recognition for the pc computer indicated growth and acceptance in the industry.
Both Allen and Gates return home to Bellevue, Washington and announce plans to open offices in their home town, thus becoming the first microcomputer software company in the Northwest.
Technical Details of both Linux and Windows OS’s
An OS takes care of all input and output coming to a computer. It manages users, processes, memory management, printing, telecommunications, networking, and etc. The OS sends data to a disk, the printer, the screen and other peripherals connected to the computer. A computer can’t work without an OS. The OS tells the machine how to process instructions coming from input devices and software running on the computer. Therefore every computer is built different, commands for in or output will have to be treated differently. In most cases an operating system is not a gigantic nest of programs but instead a small system of programs that operate by the core or kernal. The pc computer system is so compact these small supporting programs it is easier to rewrite parts r packages of the system than to redesign an entire program.
When first created OS’s were designed to help applications interact with the computer hardware. This is the same today, the importance of the OS has risen to the point where the operating system defines the computer. The OS gives off a layer of abstraction between the user and the machine when they communicate. Users don’t see the hardware directly, but view it through the OS. This abstraction can be used to hide certain hardware details from the application and the user.
Applied software is that which is not generic but specifically for one single task machine. The software will not run on any other machine. Applications like this are SABRE, the reservation system of airlines, and defense systems. Computer Aided Software Engineering (CASE) Creating software is an expensive and time consuming process. These programs will support and in some cases replace the engineer in creating computer programs. Cad cam systems is the computer aided design &computer aided manufacturing. The electronic drawing board in a computer program the features are multiplying. Like premanufactured elements, strength calculations, emulations of how a construction will hold in earthquakes.
In Linux there has been a question that has been going back and forth now for a while, is SCSI dead for workstations? There have been many advancements in SATA and the mainstream acceptance of 10K RPM Western Digital Raptor maybe this made SCSI too expensive for what is needed in a workstation. It’s time we take a look at Linux. How does the Western Digital Raptor WD740GD compare to the three latest Ultra320 SCSI drives: the Seagate Cheetah 10K.7, Seagate Cheetah 15K.3, and Seagate Cheetah 15K.4. This section covers the technology of the drives, acoustics, heat, size, and performance.
Lets take a look at the latest generation of the Seagate 10K Cheetah line and 15K Cheetah line. We will also be taking an in depth look at the latest 10K SATA drive from Western Digital the 74GB WD740GD. Starting with the Western Digital Raptor, WD pushes this drive as the low cost answer to SCSI. On their website, they like to show off the drives 1,200,000 hours MTBF(Mean Time Between Failure) which matches the last generation MTBF of the Seagate Cheetah 15K.3 and is very close to the reliability rating of today’s Cheetahs.
In Linux’s datasheet or newsletter, they also mention that the Cheetah drive is designed for “high performance around the clock usage.” Both the Cheetah and the Western Digital Raptor drives have the same amount of cache memory. When you are speaking of operations in a multi-tasking/multi-user environment, the benefit of various queuing techniques is an advantage. All Ultra 320 SCSI drives support what is called Native Command Queuing or NCQ. This technique is where all commands sent to the disk drive can be queued up and reordered in the most efficient order. This stops the drive from having to request service on only one side of the disk, then going to the other side of the disk serving another request, in order to return for the next request.. While some of the SATA drives do support NCQ, the Raptor does not. The Raptor does have another form of queuing called Tagged Command Queuing or TCQ. This method is not as effective as NCQ and requires support in both the drive and host controller. From what they have been able to determine, TCQ support is sparse, even under Windows.
The SATA drive has itself backed up on their durability claim by stating their use of fluid dynamic bearings in their drives. The fluid dynamic bearings replace ball bearings to cut down on drive wear and tear and decrease operating noise.
Microsoft Windows XP technologies make it easy to enjoy games, music, and movies in addition to creating movies and enhancing digital photo’s. Direct X 9.0 technology drives high speed multimedia and various games on the PC. DirectX provides the exciting graphics, sound, music, and three dimensional animation that bring games to life. Direct X is also the link that allows software engineers to develop a game that is high speed and multimedia driven for your PC. Direct X was introduced in 1995 and it’s popularity soared as multimedia application development reached new heights. Today Direct X has progressed to an Application Programming Interface (API) and being applied into Microsoft Windows Operating Systems. This way software developers can access hardware features without having to write hardware code.
Some of the features of the windows media playerb 9 series with smart jukebox gives users more control over their music. With easy cd transfer to the computer, cd burning and compatibility is available on portable players. Users can also discover more with services that have premium entertainment. Windows media player 9 seriers works well with windows xp using the built in digital media features and delivers a state-of- the- art experience.
When Windows Millenium Edition 2000 came out of stores it was specifically designed for home users. It had the first Microsoft version of a video editing product. Movie Maker is used to capture and organize and edit video clips, and then export them for PC or web playback. Movie maker 2, released in 2003, adds new movie making transitions, jazzy titles, and neat special effects. Based on Microsoft Direct Show and Windows Media technologies, Movie Maker was originally included only with Windows Millenium Edition. Now Movie Maker 2 is available for Windows XP Home Edition and Windows XP Professional.
With the release of Windows XP in 2001 came Windows Messenger, bringing instant messaging to users across the internet. Users communicate using Text messages in real time in Windows Messenger. Real time messaging with video conferencing has been available for a long time before now. The first communication tool provided by Windows Messenger used integrated, easy to use text chat, voice and video communication, and data collaboration.
Linux is being developed and thus is freely redistributable in code form. Linux is available and developed over the internet. Many of the engineers who took part in producing it are from over seas and have never meet one another. This operating system is at a source level code and is on a large scale that has led the way to it becoming an featureful and stable system.
Eric Raymond has written a popular essay on the development of Linux entitled The Cathedral. and the bazaar. He describes the way the Linux kernal uses a Bazaar approach that has the code released quickly and very often, and that this requires input that has provided improvement to the system. This Bazaar approach is reported to the Cathedral approach used by other systems like GNU Emacs core. The Cathedral approach is characterized in bringing a more beautiful code that has been released, but unfortunately it is released far less often. A poor opportunity for people outside the group who can not contribute to the process.
Some of the high-lights and success of the Bazaar projects do not include the opening the code for everyone to observe, at the design level of the Bazaar. On the same token the Cathedral approach is widely viewed by everyone and is appropriate. Once debugging the code is executed, it is necessary to open the Bazaar to have everyone find different errors involving the code. If they can fix the code this a great effort and help to the coders.
Advantages and Disadvantages of the two OS’s
The writer of this Linux OS web page Chris Browne, describes the way that Linux efforts are distributed and some of the advantages and disadvantages of the Linux OS. The Linux OS comes with some experimental versions such as the 2.5. x series where version numbers go steadily upwards every week. The stable version changes only when bugs are detected in the system and the bugs must be fixed in the experimental series, and this occurrence does not change very often. Linux users know that this happens, and they work to resolve the bugs.
It is not guaranteed that all users will immediately fix their problems with the systems if they are not being affected (or don’t notice they are affected) by problems, there are fixes quickly available, sometimes distributed across the internet after a few hours of diagnosis. For Linux fixes are available more quickly than commercial vendors like Microsoft, HP, and IBM usually this diagnosis is before they even know there’s a problem. This acknowledgement is in contrast to other companies behavior, Bill Gates claims in his press releases Microsoft code has no bugs. This seems to mean that there are no bugs that Microsoft cares to fix.
Microsoft came to the conclusion that the majority of bugs detected in their systems are present because users don’t use their software correctly. The problems that remain for Microsoft are few in number and are caused by actual errors. There is remaining work to get the stable Linux system, with configured Linux kernels that should and do have suitably configured software on top of the workload the systems have to run for hundreds of days without rebooting the computers. Some of the general public as well as computer professionals like engineers and technicians complain that Linux is always changing. Chris says that “effort and interest of the Linux kernal will stop when people want to stop building and enhancing the Linux kernal.” As long as new technology and devices like the video cards are being constructed and people interested in Linux keep coming up with new improvements for Linux, work on Linux OS will progress.
The disadvantage of the Linux OS is that it may end because of there being a better platform for kernal hacking, or because Linux in the future will be so displaced that it becomes unmanageable. This has not happened yet but many researchers say that in the future of Linux, with various plans for attaining services to the consumer or business, Linux is moving away from the base kernal and into user space which creates less room for data and information. The announcement of a Debian Hurd effort suggests an alternative to the problem of kernal hacking. The Hurd kernal, which runs and is sent as a set of processes on top a microkernal such as MACH, may provide a system for those people that are not satisfied with changes to the linux kernal. Mach has a “message passing” abstraction that allows the OS to be created as a set of components that will work in conjunction with one another.
Competetive, Collaborative Efforts
To start this section I’ll tell about the beginning of the personal computer and it’s roots with IBM. Vertically integrated proprietary de facto standards architectures were the norm for the first three decades of the postwar computer industry. Each computer manufacturer made most if not all of its technology internally, and sold that technology as part of an integrated computer. This systems era was ascendant from IBM’s 1964 introduction of its System 360 until the release of the 1981, personal computer from IBM. This was challenged by two different approaches. One was the fragmentation of proprietary standards in the PC industry between different suppliers, which led Microsoft and Intel to seek industry wide dominance for their proprietary component of the overall system architecture, making what Moschella (1997) terms the “PC era” (1964-1981). The second was a movement by users and second tier producers to cvonstruct industrywide “open” systems, in which the standard was not owned by a single firm.
The adoption of the Linux system in the late 1990s was a response to these earlier approaches. Linux was the most commercially accepted example of a new wave of “open source” software, the software and the source code are freely distributed to use and modify. The advantages of Linux in contrast to the proprietary PC standards, particularly software standards controlled by Microsoft. Product compatibility standards have typically been considered using a simple unidemensional typology, bifurcated between “compatible” and “incompatible.” Further more, to illuminate differences between proprietary and open standards strategies, Gabel’s (1987) multi-dimensional classification attribute, with each dimension assuming one of several (discrete) levels:
“multivintage” compatibility between successive generations of a product:
“product line” compatibility, providing interoperability across the breadth of the company’s
product line-as Microsoft has with its Windows CE, 95/98/ME, and NT/2000 product families.
“multivendors” compatibility, i.e. compatibility of products between competing producers.
The first successful multi-vendor operating system was Unix, developed by a computer science research group at Bell Telephone Laboratories (BTL) in New Jersey beginning in 1969. As with the earlier Multics research project between MIT, BTL and mainframe computer maker General Electric, Unix was a multi-user time-shared OS designed as a research project by programmers for their personal use. Other characteristics key to Unix’s success reflected path dependencies by its developers and early users( Salus 1994):
AT&T was forbidden by its 1956 consent decree from being in the computer business, so it did not sell the OS commercially. After publishing research papers, Bell Labs was flooded with requests from university computer science departments, who received user licenses and source code but a lack of support. Along cam budget constraints that limited BTL researchers to DEC minicomputers opposed to large mainframe computers, Unix was simpler and more efficient than its Multics predecessor, based on the simplified C programming language rather than the more widely used PL/I. Although originally developed DEC minicomputers, Unix was converted to run on other models by users who found programmer time less expensive than buying a supported model, thus setting the stage for it to become a hardware-independent OS.
Maybe one of the most important developments was the licensing of UNIX by the U.C. Berkeley Computer Science Department in 1973. The Berkeley group issued its own releases from 1977 to 1994, with much of its funding provided by the Defense Advanced Research Projects Agency (DARPA). The result of the Berkeley development included (Garud and Kumaraswamy 1993; Salus 1994) :
The first Unix version to support TCP/IP, later the standard protocols of the internet;
Academic adoption of BSD Unix as the preferred OS by many computer science departments throughout the world;
Commercial spread of BSD -derived Unix through Sun Microsystems, cofounded by former BSD programmer Bill Joy;
As they evolved their versions of Unix, fragmentation of Unix developers and adopters into rival “BSD” and “AT&T” camps.
AT&T Unix provided a multivendor standard which, when coupled with the BSD advancements, helped spur the adoption of networked computing. Helped by Sun, whose slogan is “the network is the computer,” Unix rapidly gained acceptance during the 1980s as the preferred OS for networked engineering workstations (Garud and Kumaraswamy 1993). At the same time, it became a true multivendor standard as minicomputer producers with a small amount of customers, weak R&D and immature OS licensed Unix from AT&T. The main exceptions to the Unix push were the early leaders in workstations (Apollo) and minicomputers (DEC), who used their proprietary OS as a source of competitive advantage, and were the last to switch to Unix in their respective segments.
Some of the advocates from the two producers formed a number of trade associations to promote Unix and related operating systems. In doing so fueled the adoption and standardization of Unix, they hoped to increase the amount of application software to compete with sponsored, proprietary architectures(Gabel 1987; Grindley 1995). These two groups promoted these under the rubric “open systems”; the editors of a book series on such systems summarized their goals as follows:
Open systems allow users to move their applications between systems easily; purchasing decisions can be made on the basis of cost-performance ratio and vendor support, rather than on systems which run a users application suite (Salus 1994: v).
Despite these goals, the Unix community spent the 1980s and early 1990s fragmented into AT&T and Berkeley warring factions, each of which sought control of the OS API’s to maximize the software available for their versions. Each faction had its own adherents. To avoid paying old earlier mainframe switching costs, U.S. Department of Defense procurement decisions began to favor Unix over proprietary systems. As AT&T formalized its System V Interface Definition and encouraged hardware makers to adopt System V, it became the multivendor standard required by DoD procurements
BSD group was only developed for DEC minicomputers, its Unix variant was not multivendor and less attractive and appealing for DoD procurements. The numerous innovations of the BSD group in terms of usability, software development tools and networking made it more attractive to university computer scientists for their own research and teaching, making it the minicomputer OS preferred by computer science departments in the U.S., Europe and Japan (Salus 1994). The divergent innovation meant that the two major Unix variants differed in terms of internal structure, user commands and application programming interfaces (APIs). It was the latter difference that most seriously affected computer buyers, as custom software developed for one type of Unix could not directly be recompiled on the other, adding switching costs between the two systems. Also, both the modem-based and DARPA networking facilitated the distribution of user donated source code libraries, that were free but often required site-specific custom programming if the Unix API’s at the users site differed from those of faced by the original contributor.
Microsoft Windows continues to invest in products based on the Itanium processor family, and the Itanium Solutions Alliance will further this investment by helping growth of the ecosystem of applications and solutions available on Windows platform and SQL Server 2005,” said Bob Kelly, general manager, Windows infrastructure, Microsoft Corp. “We look forward to working with the members of the Itanium Solutions Alliance to help IT managers transition from RISC-based Unix servers to Itanium based systems running on the Windows platform.”
More Stories
Where to Find the Best Royalty Free house music!
Where to Find the Best Royalty Free country music Instrumentals!
Where to Find the Best Royalty Free Hip Hop Instrumentals!