PERSONAL AND SERVER COMPUTERS HAVING MICROCHIPS WITH MULTIPLE PROCESSING UNITS AND INTERNAL FIREWALL
This invention generally relates to one or more computer networks having computers like personal computers or network servers with microprocessors linked by broadband transmission means and having hardware, software, firmware, and other means such that at least one parallel processing operation occurs that involve at least two computers in the network. More particularly, this invention relates to one or more large networks composed of smaller networks and large numbers of computers connected, like the Internet, wherein more than one separate parallel processing operation involving more than one different set of computers occurs simultaneously and wherein ongoing processing linkages can be established between virtually any microprocessors of separate computers connected to the network. Still more particularly, this invention relates to business arrangements enabling the shared used of network microprocessors for parallel and other processing, wherein personal computer owners provide microprocessor processing power to a network, preferably for parallel processing, in exchange for network linkage to other personal and other computers supplied by network providers, including linkage to other microprocessors for parallel or other processing; the basis of the exchange between owners and providers being whatever terms to which the parties agree, subject to governing laws, regulations, or rules, including payment from either party to the other based on periodic measurement of net use or provision of processing power.
This invention generally relates to one or more computer networks having computers like personal computers or network computers such as servers with microprocessors preferably linked by broadband transmission means and having hardware, software, firmware, and other means such that at least two parallel processing operations occur that involve at least two sets of computers in the network or in networks connected together, a form of metacomputing. More particularly, this invention relates to one or more large networks composed of smaller networks and large numbers of computers connected, like the Internet, wherein more than one separate parallel or massively parallel processing operation involving more than one different set of computers occurs simultaneously. Even more particularly, this invention relates to one or more such networks wherein more than one (or a very large number of) parallel or massively parallel microprocessing processing operations occur separately or in an interrelated fashion; and wherein ongoing network processing linkages can be established between virtually any microprocessors of separate computers connected to the network.
Still more particularly, this invention relates generally to a network structure or architecture that enables the shared used of network microprocessors for parallel processing, including massive parallel processing, and other shared processing such as multitasking, wherein personal computer owners provide microprocessor processing power to a network, preferably for parallel or massively parallel processing or multitasking, in exchange for network linkage to other personal and other computers supplied by network providers such as Internet Service Providers (ISP's), including linkage to other microprocessors for parallel or other processing such as multitasking. The financial basis of the shared use between owners and providers would be whatever terms to which the parties agree, subject to governing laws, regulations, or rules, including payment from either party to the other based on periodic measurement of net use or provision of processing power or preferably involving no payment, with the network system (software, hardware, etc) providing an essentially equivalent usage of computing resources by both users and providers (since any network computer operated by either entity can potentially be both a user and provider of computing resources alternately (or even simultaneously, assuming multitasking), with potentially an override option by a user (exercised on the basis, for example, of user profile or user's credit line or through relatively instant payment).
Finally, this invention relates to a network system architecture including hardware and software that will provide use of the Internet or its future equivalents or successors (and most other networks) without cost to most users of personal computers or most other computers, while also providing those users (and all other users, including of supercomputers) with computer processing performance that will at least double every 18 months through metacomputing means. The metacomputing performance increase provided by this new MetaInternet (or Metanet for short) will be in addition to all other performance increases, such as those already anticipated by Moore's Law.
By way of background, the computer industry has been governed over the last 30 years by Moore's Law, which holds that the circuitry of computer chips has been shrunk substantially each year, yielding a new generation of chips every 18 months with twice as many transistors, so that microprocessor computing power is effectively doubled every year and a half.
The long term trend in computer chip miniaturization is projected to continue unabated over the next few decades. For example, slightly more than a decade ago a 16 kilobit DRAM memory chip (storing 16,000 data bits) was typical; the current standard 16 megabit chip (16,000,000 data bits), introduced in 1993, is a thousand times larger. Industry projections are for 16 gigabit memory chips (16,000,000,000 data bits) to be introduced in 2008 and 64 gigabit chips in 2011, with 16 terabit chips (16,000,000,000,000 data bits) conceivable by the mid-to-late 2020's. This is a thousand-fold increase regularly every fifteen years. Hard drive speed and capacity are also growing at a spectacular rate, even higher than that of semiconductor microchips in recent years.
Similarly regular and enormous improvements are anticipated to continue in microprocessor computing speeds, whether measured in simple clock speed or MIPS (millions of instructions for second) or numbers of transistors per chip. For example, performance has improved by four or five times every three years since Intel launched its X86 family of microprocessors used in the currently dominant “Wintel” standard personal computers. The initial Intel Pentium Pro microprocessor was introduced in 1995 and is a thousand times faster than the first IBM standard PC microprocessor, the Intel 8088, which was introduced in 1979. The fastest of current microprocessors like Digital Equipment Corp.'s Alpha chip is faster than the processor in the original Cray Y-MP supercomputer.
Both microprocessors and software (and firmware and other components) are also evolving from 8 bit and 16 bit systems into 32 bit systems that are becoming the standard today, with some 64 bit systems like the DEC Alpha already introduced and more coming, such as Intel's Merced microprocessor in 1999, with future increases to 128 bit likely some later.
A second major development trend in the past decade or so has been the rise of parallel processing, a computer architecture utilizing more than one CPU microprocessor (often many more, even thousands of relatively simple microprocessors, for massively parallel processing) linked together into a single computer with new operating systems having modifications that allow such an approach. The field of supercomputing has been taken over by this approach, including designs utilizing many identical standard personal computer microprocessors.
Hardware, firmware, software and other components specific to parallel processing are in a relatively early stage of development compared to that for single processor computing, and therefore much further design and development is expected in the future to better maximize the computing capacity made possible by parallel processing. One potential benefit that will likely be available soon is system architecture that does not rely on the multiple microprocessors having to share memory, thereby allowing more independent operation of those microprocessors, each with their own discrete memory, like current personal computers, workstations and most other computer systems architecture; for unconstrained operation, each individual microprocessor must have rapid access to sufficient memory.
Several models of personal computers are now available with more than one microprocessor. It seems inevitable that in the future personal computers, broadly defined to include versions not currently in use, will also employ parallel computing utilizing multiple microprocessors or massively parallel computing with very large numbers of microprocessors. Future designs, such Intel's Merced chip, will have a significant number of parallel processors on a single microprocessor chip.
A form of parallel processing called superscalar processing is also being employed within microprocessor design itself. The current generation of microprocessors such at the Intel Pentium have more than one data path within the microprocessor in which data can be processed, with two to three paths being typical now and as many as eight in 1998 in IBM's new Power 3 microprocessor chip.
The third major development trend is the increasing size of bandwidth, which is a measure of communications, power or transmission speed (in terms of units of data per second) between computers connected by a network. Before now, the local area networks and telephone lines typically linking computers including personal computers have operated at speeds much lower than the processing speeds of a personal computer. For example, a typical 1997 Intel Pentium operates at 100 MIPS (millions of instructions per second), whereas the most common current Ethernet connecting PC's is roughly 10 times slower at 10 megabits per second (Mbps), although some Ethernet connections are now 100 Mbps) and telephone lines are very much slower, the highest typical speed in 1998 being about 56 kilobits (reached only during downloads, however).
Now, however, the situation is expected to change dramatically, with bandwidth or transmission speed being anticipated to expand from 5 to 100 times as fast as the rise of microprocessor speeds, due to the use of coaxial cable, wireless, and especially fiber optic cable, instead of old telephone twisted pair lines. Telecommunication providers are now making available fiber connections supporting bandwidth of 40 gigabits and higher.
Technical improvements are expected in the near term which will make it possible to carry over 2 gigahertz (billions of cycles per second) on each of 700 wavelength streams, adding up to more than 1,400 gigahertz on every single fiber thread. Experts currently estimate that the bandwidth of optical fiber has been utilized one million times less fully than the bandwidth of coaxial or twisted pair copper lines. Within a decade, 10,000 wavelength streams per fiber are expected and 20-80 wavelengths on a single fiber is already commercially available.
Other network connection developments such as asynchronous transfer mode (ATM) and digital signal processors, which are improving their price/performance tenfold every two years, are also supporting the rapid increase in bandwidth. The increase in bandwidth reduces the need for switching and switching speed will be greatly enhanced when practical optical switches are introduced in the fairly near future, potentially reducing costs substantially.
The result of this huge bandwidth increase will be extraordinary: within just a few years it will be technically possible to connect virtually any computer to a network at a speed that equals or exceeds the computer's own internal system bus speed, even as that bus speed itself is increasing significantly. The system bus of a computer is its internal network connecting many or most of its internal components such as microprocessor, random access memory (RAM), hard-drive, modem, floppy drive, and CD-ROM; for recent personal computers it has been only about 40 megabits per second, but is up to 133 megabits per second on Intel's Pentium PCI bus in 1995. IBM's 1998 Power3 microprocessor chip has a system bus of 1.6 gigabits per second.
Despite these tremendous improvements anticipated in the future, the unfortunate present reality is that a typical personal computer (PC) is already so fast that its microprocessor is essentially idle during most of the time the PC is in actual use and that operating time itself is but a small fraction of those days the PC is even in any use at all. The reality is that nearly all PC's are essentially idle during roughly all of their useful life. A realistic estimate is that its microprocessor is in an idle state 99.9% of the time (disregarding current unnecessary microprocessor busywork like executing screen saver programs, which have been made essentially obsolete by power-saving CRT monitor technology, which is now standard in the PC industry).
Given the fact that the reliability of PC's is so exceptionally high now, with the mean time to failure of all components typically several hundred thousand hours or more, the huge idle time of PC's represents a total loss; given the high capital and operating costs of PC's, the economic loss is very high. PC idle time does not in effect store a PC, saving it for future use, since the principle limiting factor to continued use of today's PC's is obsolescence, not equipment failure from use.
Moreover, there is growing concern that Moore's Law, which as noted above holds that the constant miniaturization of circuits results in a doubling of computing power every 18 months, cannot continue to hold true much longer. Indeed, Moore's Law may now be nearing its limits for silicon-based devices, perhaps by as early as 2004, and no new technologies have yet emerged that currently seem with reasonable certainty to have the potential for development to a practical level by then, although many recent advances have the potential to maintain Moore's Law.
SUMMARY OF THE INVENTIONHowever, the confluence of all three of the established major trends summarized above—supercomputer-like personal computers, the spread of parallel processing using personal computer microprocessors (particularly massively parallel processing), and the enormous increase in network communications bandwidth—will make possible in the near future a surprising solution to the hugely excessive idleness problem of personal computers (and to the problematic possible end of Moore's Law), with very high potential economic savings.
The solution is use those mostly idle PC's (or their equivalents or successors) to build a parallel or massively parallel processing computer utilizing a very large network like the Internet or, more specifically, like the World Wide Web (WWW), or their equivalents or eventual successors like the MetaInternet (and including Internet II, which is under development now and which will utilize much broader bandwidth and will coexist with the Internet, the structure of which is in ever constant hardware and software upgrade and including the SuperInternet based on essentially all optical fiber transmission) with extremely broad bandwidth connections and virtually unlimited data transmission speed.
The prime characteristic of the Internet is of course the very large number of computers of all sorts already linked to it, with the future potential for effectively universal connection; it is a network of networks of computers that provides nearly unrestricted access (other than cost) worldwide. The soon-to-be available very broad bandwidth of network communications can be used to link personal computers externally in a manner at least equivalent to the faster internal system buses of the personal computers, so that no external processing constraint will be imposed on linked personal computers by data input or output, or throughput; the speed of the microprocessor itself will be the only processing constraint of the system, other than the internal system bus design.
This will make external parallel processing possible, including massively parallel processing, in a manner paralleling more conventional internal parallel processing, call superscalar processing.
The World Wide Web (or its equivalents or successors) would thereby have the potential to be transformed into a huge virtual massively parallel processing computer or computers, with a unique potential through its established hyperlinks connections to operate in a manner at least somewhat like a human neural network or neural networks, since the speed of transmission in the linkages would be so great that any linkage between two microprocessors would be virtually equivalent to direct, physically close connections between those microprocessors.
With further development, digital signal processor-type microprocessors and/or analogue microprocessors may be particularly advantageous for this approach, either alone or in conjunction with conventional microprocessors and/or those new microprocessors described in this application. Networks with WWW-type hyperlinks incorporating digital signal processor-type microprocessor (or successors or equivalents) could operate separately from networks of conventional microprocessors (or successors or equivalents) or with one or more connections between such differing networks or with relatively complete integration between such differing networks. Simultaneous operation across the same network connection structure should be possible, employing non-interfering transmission links.
Such extremely broad bandwidth networks of computers will enable every PC to be fully utilized or nearly so. Because of the extraordinary extent to which existing PC's are currently idle, at optimal performance this new system will potentially result in a thousand-fold increase in computer power available to each and every PC user (and any other user); and, on demand, almost any desired level of increased power, limited mostly by the increased cost, which however would be relatively far less that possible from any other conceivable computer network configuration. This revolutionary increase is on top of the extremely rapid, but evolutionary increases already occurring in the computer/network industry discussed above.
The metacomputing hardware and software means of the MetaInternet will provide performance increases that will likely at least double every eighteen months based on the doubling of personal computers shared in a typical parallel processing operation by a standard PC user, starting first with at least 2 PC's, then about 4, about 8, about 16, about 32, about 64, about 128, about 256, and about 512. After about fifteen years, each standard PC user will likely be able to use about 1024 personal computers for parallel processing or any other shared computing use, while generally using the Internet or its successors like the MetaInternet for free. At the other end of the performance spectrum, supercomputers will experience a similar performance increase generally, but ultimately the performance increase is limited primarily by cost of adding temporary network linkages to available PC's, so there is definite potential for a quantum leap in supercomputer performance.
Network computer systems as described above offer almost limitless flexibility due to the abundant supply of heretofore idle connected microprocessors. This advantage would allow “tightly coupled” computing problems (which normally are difficult to process in parallel) to be solved without knowing in advance (as is now necessary in relatively massively parallel processing) how many processors are available, what they are and their connection characteristics. A minimum number of equivalent processors (with equivalent other specs) can be easily found nearby in a massive network like the Internet and assigned within the network from those multitudes available nearby. Moreover, the number of microprocessors used can be almost completely flexible, depending on the complexity of the problem, and limited only by cost. The current problem of time delay will be solved largely by the widespread introduction of extremely broad bandwidth connections between computers processing in parallel.
The new network computer will utilize PC's as providers of computing power to the network, not just users of network services. These connections between network and personal computer are enabled by a new form of computer/network financial structure that is rooted on the fact that economic resources being provided the network by PC owners (or leaser) are similar in value to those being provided by the network provider providing connectivity.
Unlike existing one way functional relationships between network providers such as internet service providers (often currently utilizing telecommunications networks for connectivity) and PC users, wherein the network provider provides access to a network like the Internet for a fee (much like cable TV services), this new relationship would recognize that the PC user is also providing the network access to the user's PC for parallel computing use, which has a similar value. The PC thus both provides and uses services on the network, alternatively or potentially even virtually simultaneously in a multitasking mode.
This new network would operate with a structural relationship that would be roughly like that which presently can exist between an electrical power utility and a small independent power generator connected to the utility, wherein electrical power can flow in either between utility and independent direction depending on the operating decisions of both parties and at any particular point in time each party is in either a debt or credit position relative to the other based on the net direction of that flow for a given period, and is billed accordingly In the increasingly deregulated electrical power industry, electrical power (both its creation and transmission) is becoming a commodity bought and sold in a competitive marketplace that crosses traditional borders. With the structural relationship proposed here for the new network, parallel free market structures should develop over time in a new computer power industry dominated by networks of personal computers in all their current and future forms providing shared processing.
For this new network and its structural relationships, a network provider is defined in the broadest possible way as any entity (corporation or other business, government, not-for-profit, cooperative, consortium, committee, association, community, or other organization or individual) that provides personal computer users (very broadly defined below) with initial and continuing connection hardware and/or software and/or firmware and/or other components and/or services to any network, such as the Internet and Internet II or WWW or their present or future equivalents, coexistors or successors, like the herein proposed MetaInternet, including any of the current types of Internet access providers (ISP's), including telecommunication companies, television cable or broadcast companies, electrical power companies, satellite communications companies, or their present or future equivalents, coexistors or successors. The connection means used in the networks of the network providers, including between personal computers or equivalents or successors, would preferably be very broad bandwidth, by such means as fiber optic cable or wireless for example, but not excluding any other means, including television coaxial cable and telephone twisted pair, as well as associated gateways, bridges, routers, and switches with all associated hardware and/or software and/or firmware and/or other components and their present or future equivalents or successors. The computers used by the providers include any computers, including mainframes, minicomputers, servers, and personal computers, and associated their associated hardware and/or software and/or firmware and/or other components, and their present or future equivalents or successors.
Other levels of network control beyond the network provider will also exist to control any aspect of the network structure and function, any one of which levels may or may not control and interact directly with the PC user. For example, at least one level of network control like the World Wide Web Consortium (W3C) or Internet Society (ISOC) or other ad hoc industry consortia) would establish and ensure compliance with any prescribed network standards and/or protocols and/or industry standard agreements for any hardware and/or software and/or firmware and/or other component connected to the network. Under the consensus control of these consortia/societies, other levels of network control would deal with administration and operation of the network. These other levels of network control might be constituted by any network entity, including those defined immediately above for network providers.
The principal defining characteristic of the network provided being communication connections (including hardware and/or software and/or firmware and/or other component) of any form, including electromagnetic (such as light and radio or microwaves) and electrochemical (and not excluding biochemical or biological), between PC users, with connection (either directly or indirectly) to the largest number of users possible being highly advantageous, such as networks like the Internet (and Internet II and SuperInternet) and WWW and equivalents and successors, like the MetaInternet. Multiple levels of such networks will likely coexist with different technical capabilities, like Internet and Internet II, but would have interconnection and therefore would communicate freely between levels, for such standard network functions as electronic mail.
And a personal computer (PC) user is defined in the broadest possible way as any individual or other entity using a personal computer, which is defined as any computer, digital or analog or neural, particularly including microprocessor-based personal computers having one or more microprocessors (each including one or more parallel processors) in their general current form (hardware and/or software and/or firmware and/or any other component) and their present and future equivalents or successors, such as workstations, network computers, handheld personal digital assistants, personal communicators such as telephones and pagers, wearable computers, digital signal processors, neural-based computers (including PC's), entertainment devices such as televisions, video tape recorders, videocams, compact or digital video disk (CD or DVD) player/recorders, radios and cameras, other household electronic devices, business electronic devices such as printers, copiers, fax machines, automobile or other transportation equipment devices, and other current or successor devices incorporating one or more microprocessors (or functional or structural equivalents), especially those used directly by individuals, utilizing one or more microprocessors, made of inorganic compounds such as silicon and/or other inorganic or organic compounds; current and future forms of mainframe computers, minicomputers, microcomputers, and even supercomputers are also be included. Such personal computers as defined above have owners or Teasers, which may or may not be the same as the computer users. Continuous connection of computers to the network, such as the Internet, WWW, or equivalents or successors, is preferred.
Parallel processing is defined as one form of shared processing as involving two or more microprocessors involved in solving the same computational problem or other task. Massively parallel microprocessor processing involves large numbers of microprocessors. In today's technology, massive parallel processing can probably be considered to be about 64 microprocessors (referred to in this context as nodes) and over 7,000 nodes have been successfully tested in an Intel supercomputer design using PC microprocessors (Pentium Pros). It is anticipated that continued software improvements will make possible a much larger number of nodes, very possibly limited only by the number of microprocessors available for use on a given network, even an extraordinarily large one like the Internet or its equivalents and/or successors, like the MetaInternet.
Broadband wavelength or broad bandwidth network transmission is defined here to mean a transmission speed (usually measured in bits per second) that is at least high enough (or roughly at least equivalent to the internal clock speed of the microprocessor or microprocessors times the number of microprocessor channels equaling instructions per second or operations per second or calculations per second) so that the processing input and output of the microprocessor is substantially unrestricted, particularly including at peak processing levels, by the bandwidth of the network connections between microprocessors that are performing some form of parallel processing, particularly including massive parallel processing. Since this definition is dependent on microprocessor speed, it will increase as microprocessor speeds increase. A rough example might be a current 100 MIPS (millions instructions per second) microprocessor, for which a broad bandwidth connection would be greater than 100 megabits per second (Mbps); this is a very rough approximation. However, a preferred connection means referenced above is fiber optic cable, which currently already provides multiple gigabit bandwidth on single fiber thread and will improve significantly in the future, so the preferred general use of fiber optic cable virtually assures broad bandwidth for data transmission that is far greater than microprocessor speed to provide data to be transmitted. The connection means to provide broad bandwidth transmission can be either wired or wireless, with wireless generally preferred for mobile personal computers (or equivalents or successors) and as otherwise indicated below. Wireless connection bandwidth is also increasing rapidly and can be considered to offer essentially the same principal benefit as fiber optic cable: external data transmission speed in a network that far exceeds internal data processing speed in any computer.
The financial basis of the shared use between owners/leasers and providers would be whatever terms to which the parties agree, subject to governing laws, regulations, or rules, including payment from either party to the other based on periodic measurement of net use or provision of processing power.
In one embodiment, as shown in
In another embodiment, as shown in
Alternately, as shown in
Preferably, the network would involve no payment between users and providers with the network system (software, hardware, etc.) providing an essentially equivalent usage of computing resources by both users and providers (since any network computer operated by either entity can potentially be both a user and provider of computing resources (even simultaneously, assuming multitasking), with potentially an override option by a user (exercised on the basis, for example, of user profile or user's credit line or through relatively instant payment).
Preferably, as shown in
A specific category of PC user based, for example, on specific microprocessor hardware owned or leased, might have access to a set maximum number of parallel PC's or microprocessors, with smaller or basic users generally having less access and vice versa. Specific categories of users might also have different priorities for the execution of their processing by the network. A very wide range of specific structural forms between user and provider are possible, both conventional and new, based on unique features of the new network computer system of shared processing resources.
For example, in the simplest case, in an initial system embodiment, as shown in
Preferably, for most standard PC users (including present and future equivalents and successors), connection to the Internet (or present or future equivalents or successors like the MetaInternet) would be at no cost to PC users, since in exchange for such Internet access the PC users would generally make their PC, when idle, available to the network for shared processing. Preferably, then, competition between Internet Service Providers (including present and future equivalents and successors) for PC user customers would be over such factors as the convenience and quality of the access service provided and of shared processing provided at no addition cost to standard PC users, or on such factors as the level of shared processing in terms, for example of number of slave PC's assigned on a standard basis to a master PC. The ISP's would also compete for parallel processing operations, from inside or outside the ISP Networks, to conduct over their networks.
In addition, as shown in
Or, more simply, as shown in
In a preferred embodiment, as shown in
Also, in another embodiment, as shown in
In one embodiment, as shown in
One of the primary capabilities of the Internet (or Internet II or successor, like the MetaInternet) or WWW network computer would be to facilitate searches by the PC user or other user. As shown in
As a typical example, a single PC user might need 1,000 minutes of search time to find what is requested, whereas the network computer, using multiple PC processors, might be able to complete the search in 100 minutes using 10 processors, or 10 minutes using 100 processors or 1 minute using 1,000 processors (or even 1 second using 60,000 processors); assuming performance transparency, which should be achievable, at least over time. The network's external parallel processing is completely scalable, with virtually no theoretical limit.
The above examples also illustrates a tremendous potential benefit of network parallel processing. The same amount of network resources, 60,000 processor seconds, was expended in each of the equivalent examples. But by using relatively large multiples of processors, the network can provide the user with relatively immediate response with no difference in cost. (or relatively little difference)—a major benefit. In effect, each PC user linked to the network providing external parallel processing becomes, in effect, a virtual supercomputer! As discussed below, supercomputers would experience a similar quantum leap in performance by employing a thousand-fold (or more) increase in microprocessors above current levels.
Such power will likely be required for any effective searches in the World Wide Web (WWW). WWW is currently growing at a rate such that it is doubling every year, so that searching for information within the WWW will become geometrically more difficult in future years, particularly a decade hence, and it is already a very significant difficulty to find WWW sites of relevance to any given search and then to review and analyze the contents of the site.
So the capability to search with massive parallel processing will be required to be effective and will dramatically enhance the capabilities of scientific, technological and medical researchers.
Such enhanced capabilities for searching (and analysis) will also fundamentally alter the relationship of buyers and sellers of any items and/or services. For the buyer, massive parallel network processing will make it possible to find the best price, worldwide, for any product or the most highly rated product or service (for performance, reliability, etc.) within a category or the best combination of price/performance or the highest rated product for a given price point and so on. The best price for the product can include best price for shipping within specific delivery time parameters acceptable to the buyer.
For the seller, such parallel processing will drastically enhance the search, worldwide, for customers potentially interested in a given product or service, providing very specific targets for advertisement. Sellers, even producers, will be able to know their customers directly and interact with them directly for feedback on specific products and services to better assess customer satisfaction and survey for new product development.
Similarly, the vastly increased capability provided by the system's shared parallel processing will produce major improvements in complex simulations like modeling worldwide and local weather systems over time, as well as design and testing of any structure or product, from airliners and skyscrapers, to new drugs and to the use of much more sophisticated artificial intelligence (AI) in medical treatment and in sorting through and organizing for PC users the voluminous input of electronic data from “push” technologies. Improvements in games would also be evident, especially in terms of realistic simulation and interactivity.
As is clear from the examples, the Internet or WWW network computer system like the MetaInternet would potentially put into the hands of the PC user an extraordinary new level of computer power vastly greater than the most powerful supercomputer existing today. The world's total of microchips is already about 350 billion, of which about 15 billion are microprocessors of some kind (most ate fairly simple “appliance” type running wrist watches, Televisions, cameras, cars, telephones, etc). Assuming growth at its current rates, in a decade the Internet/Internet II/WWW could easily have a billion individual PC users, each providing a average total of at least 10 highly sophisticated microprocessors (assuming PC's with at least 4 microprocessors (or more, such as 16 microprocessors or 32, for example) and associated other handheld, home entertainment, and business devices with microprocessors or digital processing capability, like a digital signal processor or successor devices). That would be a global computer a decade from now made of at least 10 billion microprocessors, interconnected by electromagnetic wave means at speeds approaching the speed of light.
In addition, if the exceptionally numerous “appliance” microprocessors noted above, especially those that operate now intermittently like personal computers, are designed to the same basic consensus industry standard as parallel microprocessors for PC's (or equivalents or successors) or for PC “systems on a chip” discussed later in
To put this in context, a typical supercomputer today utilizing the latest PC microprocessors has less than a hundred. Using network linkage to all external parallel processing, a peak maximum of perhaps 1 billion microprocessors could be made available for a network supercomputer user, providing it with the power 10,000,000 times greater than would be available using today's internal parallel processing supercomputers (assuming the same microprocessor technology). Because of it's virtually limitless scalability mentioned above, resources made available by the network to the supercomputer user or PC user would be capable of varying significantly during any computing function, so that peak computing loads would be met with effectively whatever level of resources are necessary.
In summary, regarding monitoring the net provision of power between PC and network,
Also, relative to maintaining a standard cost,
Browsing functions generally include functions like those standard functions provided by current Internet browsers, such as Microsoft Explorer 3.0 or 4.0 and Netscape Navigator 3.0 or 4.0, including at least searching World Wide Web or Internet sites, exchanging E-Mail worldwide, and worldwide conferencing; an intranet network uses the same browser software, but might not include access to the Internet or WWW. Shared processing includes at least parallel processing and multitasking processing involving more than two personal computers, as defined above. The network system is entirely scalar, with any number of PC microprocessors potentially possible.
As shown in
For example, as shown in
Other typical PC hardware components such as hard drive 61, floppy drive 62, compact disk-read only memory (CD-ROM) 63, digital video disk (DVD) 64, Flash memory 65, random access memory. (RAM) 66, video or other display 67, graphics card 68, and sound card 69, as well as digital signal processor or processors, together with the software and/or firmware stored on or for them, can be located on either side of the preferred firewall 50, but such devices as the display 67, graphics card 68 and sound card 69 and those devices that both read and write and have non-volatile memory (retain data without power and generally have to be written over to erase), such as hard drive 61, Flash memory 65, floppy drive 62, read/write CD-ROM 63′ or DVD 64′ are preferred to be located on the PC user side of the firewall 50, where the master microprocessor is also located, as shown in
Alternately, any or these devices that are duplicative (or for other exceptional needs) like a second hard drive 61′ can be located on the network side of the firewall 50. RAM 66 or equivalent or successor memory, which typically is volatile (data is lost when power is interrupted), should generally be located on the network side of the firewall 50, however some can be located with the master microprocessor to facilitate its independent use.
Read-only memory devices such as most current CD drives (CD-ROM's) 63′ or DVD's (DVD-ROM) 64′ can be safely located on the network side of the firewall 50, since the data on those drives cannot be altered by network users; preemptive control of use would preferably remain with the PC user in terms of interrupting network use.
However, at least a portion of RAM is can be kept on the Master 30 microprocessor side of the firewall 50, so that the PC user can use retain the ability to use a core of user PC 1 processing capability entirely separate from any network processing; if this capability is not desired, then the master 30 microprocessor can be moved to the network side of the firewall 50 and replaced with a simpler controller on the PC 1 user side, like the master remote controller 31 discussed below and shown in
And the master microprocessor 30 might also control the use of several or all other processors 60 owned or leased by the PC user, such as home entertainment digital signal processors 70, especially if the design standards of such microprocessors in the future conforms to the requirements of network parallel processing as described above. In this general approach, the PC master processor would use the slave microprocessors or, if idle (or working on low priority, deferable processing), make them available to the network provider or others to use. Preferably, wireless connections 100 would be extensively used in home or business network systems, including use of a master remote controller 31 without (or with) microprocessing capability, with preferably broad bandwidth connections such as fiber optic cable connecting directly to at least one component such as a PC 1, shown in a slave configuration, of the home or business personal network system; that preferred connection would link the home system to the network 2 such as the Internet 3, as shown in
A PC 1 user can remotely access his networked PC 1 by using another networked master microprocessor 30 on another PC 1 and using a password or other access control means for entry to his own PC 1 master microprocessor 30 and files, as is common now in Internet and other access. Alternately, a remote user can simply carry his own files and his own master microprocessor or use another networked master microprocessor temporarily as his own.
In the simplest multi-microprocessor configuration, as shown in
Preferably, as shown in
Existing PC components with mechanical components like hard drive 61, floppy drive or other drive configured to have a removable diskette 62, CD-ROM 63 and DVD 64, which are mass storage devices with mechanical features that will likely not become an integral part of a PC “system of a chip” would preferably, of course, still be capable of connection to a single PC micro chip 90 and control by a single PC master unit 93.
In the simplest multi-processor case, as shown in
As noted in the second paragraph of the introduction to the background of the invention, in the preferred network invention, any computer can potentially be both a user and provider, alternatively—a dual mode operating capability. Consequently, any PC 1 within the network 2, preferably connected to the Internet 3 (and successors like the MetaInternet), can be temporarily a master PC 30 at one time initiating a parallel or multitasking processing request to the network 2 for execution by at least one slave PC 40, as shown in
As shown in
As noted above relative to
Similarly,
In summary, relative to the use of master/slave computers,
The preferred use of the firewall 50, as described above in
In summary, regarding the use of firewalls,
In summary, regarding the use of controllers with firewalls,
In summary, regarding the use of firewalls that can be actively configured,
It is presently contemplated that PC 1 microprocessors noted above be designed to the same basic consensus industry standard as parallel microprocessors for PC's (or equivalents or successors) as in
By 1998, manufacturing technology improvements will allow as many as 20 million transistors to fit on a single chip (with circuits as thin as 0.25 microns) and, in the next cycle, 50 million transistors using 0.18 micron circuits. Preferably, that entire computer on a chip would be linked, preferably directly, by fiber optic or other broad bandwidth connection means to the network so that the limiting factor on data throughput in the network system, or any part, is only the speed of the linked microprocessors themselves, not the transmission speed of the linkage. Such direct fiber optic linkage will obviate the need for an increasingly unweldy number of microchip connection prongs, which is currently in the one to two hundred range in the Intel Pentium series and will reach over a thousand prongs in the 1998 IBM Power3 microprocessor. One or more digital signal processors 89 located on a microprocessor, together with numerous channels and/or signal multiplexing of the fiber optic signal can substitute for a vast multitude of microchip connection prongs.
For computers that are not reduced to a single chip, it is also preferred that the internal system bus or buses of any such PC's have a transmission speed that is at least high enough that the all processing operations of the PC microprocessor or microprocessors is unrestricted (and other PC components like RAM) and that the microprocessor chip or chips are directly linked by fiber optic or other broad bandwidth connection, as with the system chip described above, so that the limiting factor on data throughput in the network system, or any part, is only the speed of the linked microprocessors themselves, not the transmission speed of the linkage.
The individual user PC's can be connected to the Internet (via an Intranet)/Internet II/WWW or successor, like the MetaInternet (or other) network by any electromagnetic means, with the speed of fiber optic cable being preferred, but hybrid systems using fiber optic cable for trunk lines and coaxial cable to individual users may be more cost effective initially, but much less preferred unless cable can be made (through hardware and/or software and/or firmware and/or other component means) to provide sufficiently broad bandwidth connections to provide unrestricted throughput by connected microprocessors. Given the speed and bandwidth of transmission of fiber optic or equivalent connections, conventional network architecture and structures should be acceptable for good system performance, making possible a virtual complete-interconnection network between users.
However, the best speed for any parallel processing operation should be obtained, all other things being equal, by utilizing the available microprocessors that are physically the closest together. Consequently, as shown previously in
Network architecture that clusters PC's together should therefore be preferred and can be constructed by wired means. However, as shown in
It would be advantagous, then, for those wireless PC connections to be PC resident and capable of communicating by wireless or wired (or mixed) means with all available PC's in the cluster or cell geographic area, both proximal and potentially out to the practical limits of the wireless transmission.
As shown in
Moreover, as shown in
As shown in
This same transponder approach also can be used between PC's 1″ connected by a wired 99 (or mixed wired/wireless) means, despite the fact that connection distances would generally be greater (since not line of sight, as is wireless), as shown in
In that case, a functionally equivalent transponder device 120 would preferably be operated in wired clusters 101 in generally the same manner as described above for PC's connected in wireless clusters 101. Networks incorporating PC's 1 connected by both wireless and wired (or mixed) means are anticipated, like the home or business network mentioned in
As shown in
The
To improve response speed in shared processing involving a significant number of slave PC's 1, a virtual potential parallel processing network for PC's 1 in a cluster 101 would preferably be established before a processing request begins. This would be accomplished by the transponder device 120 in each idle PC 1, a potential slave, broadcasting by transponder 120 its available state when it becomes idle and/or periodically afterwards, so that each potential master PC 1 in the local cluster 101 would be able to maintain relatively constantly its own directory 121 of the idle PC's 1 closest to it that are available to function as slaves. The directory 121 would contain, for example, a list of about the standard use number of slave PC's 1 for the master PC (which initially would probably be just one other PC 1″) or a higher number, preferably listed sequentially from the closest available PC to the farthest. The directory of available slave PC's 1 would be preferably updated on a relatively up to date basis, either when a change occurs in the idle state of a potential slave PC in the directory 121 or periodically.
Such ad hoc clusters 101 should be more effective by being less arbitrary geographically, since each individual PC would be effectively in the center of its own ad hoc cluster. Scaling up or down the number of microprocessors required by each PC at any given time would also be more seamless.
The complete interconnection potentially provided by such highly effective ad hoc wireless clusters is also remarkable because such clusters mimics the neural network structure of the human brain, wherein each nerve cell, called a neuron, interconnects in a very complicated way with the neurons around it. By way of comparison, the global network computer described above that is expected in a decade will have at least about 10 times as many PC s as a human brain has neurons and they will be connected by electromagnetic waves traveling at close to the speed of light, which is about 300,000 times faster than the transmission speed of human neurons (which, however, will be much closer together).
An added note: in the next decade, as individual PC's become much more sophisticated and more network oriented, compatibility issues may recede to unimportance, as all major types of PC's will be able to emulate each other and most software, particularly relative to parallel processing, will no longer be hardware specific. Nearer term it will be important to set compatible hardware, software, firmware, and other component standards to achieve substantial performance advantages for the components of the global network computer.
Until that compatibility is designed into the essential components of network system, the existing incompatibility of current components dramatically increase the difficulty involved in parallel processing across large networks. Programming languages like Java is one approach that will provide a partial means for dealing with this interim problem.
In addition, using similar configurations of existing standards, like using PC s with a specific Intel Pentium chip with other identical or nearly identical components is probably the best way in the current technology to eliminate many of the serious existing problems that could easily be designed around in the future by adopting reasonable consensus standards for specification of all system components. The potential gains to all parties with an interest far outweigh the potential costs.
The above described global network computer system has an added benefit of reducing the serious and growing problem of the nearly immediate obsolescence of computer hardware, software, firmware, and other components. Since the preferred system above is the sum of its constituent parts used in parallel processing, each specific PC component becomes less critical. As long as access to the network utilizing sufficient bandwidth is possible, then all other technical inadequacies of the user's own PC will be completely compensated for by the network's access to a multitude of technically able PC's of which the user will have temporary use.
Although the global network computer will clearly cross the geographical boundaries of nations, its operation should not be unduly bounded by inconsistent or arbitrary laws within those states. There will be considerable pressure on all nations to conform to reasonable system architecture and operational standards generally agreed upon, since the penalty of not participating in the global network computer is potentially so high as to not be politically possible anywhere.
As shown in
While the conventional approach to configuring a network of personal computers PC 1 for parallel processing is simply to string them together serially, as shown previously in
Although the
First, a great deal of complex pre-operation planning and custom tailoring-type programming at the master PC 1 level is require to establish a means for allocating portions of the operation among some number of available personal computers PC 1′.
Second, operations results coming back from personal computers PC 1′ are not synchronized, so that PC 1 frequently alternates between being idle and being overwhelmed. When the number of personal computers PC 1′ is very large, both problems can be significant.
Third, generally there is no means established for personal computers PC 1′ to communicate or cooperate with each other during operations, so sharing operational results during processing between personal computers PC 1′ is not feasible. Consequently, closely coupled problems are generally not amenable to solution by conventional parallel processing by computers.
The new hierarchical network topology shown in
Similarly, in
Like
As shown in
The offloading capability of functional roles of master and slave personal computers PC 1 (and microprocessors 30 and 40) from unavailable to available PC 1, 30 and 40 as shown in
More specifically,
In the routing of operation results shown in
Such consolidation could eliminate duplicative data from a search or other duplicative results and also serve to buffer the originating master M1 from overloading caused by many sets of results arriving in an uncoordinated fashion from what might be a large number of slave personal computers PC 1 (or microprocessors 40). Such a consolidation role for personal computers PC 1 (or microprocessors 40) would substantially reduce or eliminate the excessive custom pre-planning and synchronization problems of the conventional
Such routing splits can also vary between levels or even within the same level, as shown in
A higher level personal computer PC 1 (or microprocessor 30 or 40) such as S31 can process results as well as route them, as shown in
Together,
FIG. 16AA shows a useful embodiment in which each microprocessor 30 and 40 has its own random access memory. The design can also incorporate (or substitute) conventional shared memory (i.e. memory used by all, or some, of the microprocessors 30 or 40 of the PC 1).
As shown in
Like personal computers located in the home or office, personal computers PC 1 in automobiles (including other transportation vehicles or other conveyances) are in actual use only a very small percentage of the time, with the average dormant period of non-use totaling as much as 90 percent or more. Personal computers PC 1 are now being added to some automobiles and will likely soon become standard equipment. In addition, automobiles already have a very large number of computers onboard in the form of specialized microprocessors which are likely to become general parallel processors in future designs, as discussed earlier in this application.
Automobiles therefore form a potentially large and otherwise unused resource for massive parallel processing through the Internet 3 and other networks, as described in earlier figures. However, when idle and thus generally available for network use, automobiles lack their usual power source, the engine, which of course is then off, since it is too large to efficiently provide electrical power to onboard computers. The car engine has a controller (hardware, software or firmware or combination in the PC 1, for example) to automatically start in order to recharge the car battery when the battery is low (and before the battery is too low to start the engine) but the engine additionally needs to be controlled not to expend all available fuel automatically.
Two solutions, not mutually exclusive, to alleviate (but not solve) the problem are adding an additional car battery for network use (at least primarily) or using a single battery but adding a controller in the PC 1 for example that prevents the existing battery from being discharged to a level below that which is needed to start the automobile.
In addition, one or more solar power generating cells or cell arrays can be incorporated in an automobile's outer surface, with generally the most effective placement being on a portion of the upper horizontal surface, such as a portion of the roof, hood, or trunk. For charging the automobile battery when sunlight is not available, such as at night or in a garage, a focused or focusable light source can provide external power to the solar panel.
Alternately, a connection device such as a plug for an external electrical power source can be installed on or near the outer surface of the automobile. In addition, or independently, a connection device for an optical fiber (or other wired) external connection to the Internet 3 or other net. Alternately, a wireless receiver located near where the automobile is parked, such as in a garage, can provide connection from the automobiles personal computers. PC 1 to a network (and the Internet 3) in a home or business. This application encompasses all new apparatus and methods required to operate the above described network computer system or systems, including any associated computer or network hardware, software, or firmware (or other component), both apparatus and methods. Specifically included, but not limited to, are (in their present or future forms, equivalents, or successors): all enabling PC and network software and firmware operating systems, user interfaces and application programs; all enabling PC and network hardware design and system architecture, including all PC and other computers, network computers such as servers, microprocessors, nodes, gateways, bridges, routers, switches, and all other components; all enabling financial and legal transactions, arrangements and entities for network providers, PC users, and/or others, including purchase and sale of any items or services on the network or any other interactions or transactions between any such buyers and sellers; and all services by third parties, including to select, procure, set up, implement, integrate, operate and perform maintenance, for any or all parts of the foregoing for PC users, network providers, and/or others.
The combinations of the many elements the applicant's invention introduced in the preceding figures are shown because those embodiments are considered to be at least among the most useful possible, but many other useful combination embodiments are not shown simply because of the impossiblity of showing them all while maintaining a reasonable brevity in an unavoidably long description caused by the inherently highly interconnected nature of the inventions shown herein, which can operate all as part of one system or independently.
The forgoing embodiments meet the objectives of this invention as stated above. However, it will be clearly understood by those skilled in the art that the foregoing description has been made in terms of the preferred embodiments and that various changes and modifications may be made without departing from the scope of the present invention, which is to be defined by the appended claims.
Claims
1. A system for a network of computers, including personal computers, comprising:
- at least three personal computers;
- one said personal computer functioning as a master in a parallel processing or multi-processing operation involving at least two other said personal computers functioning as slaves to said master;
- hardware or software or firmware means for said master personal computer to subdivide said operation into two parts and to send said parts to said slaves for processing by said slaves;
- hardware or software or firmware means for said at least two slave personal computers to process said operation and send results back to said master when the operation is completed.
Type: Application
Filed: Jul 8, 2009
Publication Date: Jan 14, 2010
Inventor: Frampton Erroll ELLIS, III (Arlington, VA)
Application Number: 12/499,555
International Classification: G06F 15/16 (20060101);