The study conducted at the Massachusetts Institute of Technology covered the process of ratification of various forms of networking standards. The findings were quite amazing, especially a voting requirement that it had to go through. For instance, the research noted that the process of ratification of the “IEEE 1901 Draft Standard” that took place in April 2010 saw the first sponsor coming with an overwhelming 80% approval by voting entities. As a matter of fact, this served to set grounds for ratification of initial global power line standards of networking referred to as the “Standards Development Organization”. The successful entry of the IEEE standards of networking meant that power line networking had reached a new stage of performance. Nowadays, various stakeholders including retail vendors and Smart Grid companies can shift their product developments concerning a powerline product quite rapidly based on the foundation of worldwide standards that have been put in place to ensure interoperability and enjoy benefits of their compatibilities with several HomePlug AV based products that got an entry into the market much earlier (Villegas, 2007).
According to some specialists interviewed for the study, the process of ratification is based upon the IEEE 802.11 standards. This is basically a set of standards that have been used for the implementation of the wireless local area network also known as “WLAN”. However, this is often done in the 2.4, 3.6 and 5 GHz bands of frequency. Essentially, they are created and maintained to some extent by the IEEE Standards Committee and are provided for the creation of products of the wireless network using the brand name of “Wi-Fi”. The study has noted that the family known as 802.11 often consists of a sequence of over air techniques of modulation that applies the same protocol. However, the most popular modulation techniques are the ones associated with the 802.11b and 802.11g protocols, which are actually amendments to the original standards (Fleishman, Glenn, 2009).
The study has analyzed the current situation concerning a multicore processor and evaluated potential developments in that area. A multicore processor is defined as a single component computing system with multiple independent processors considered as units that read and execute instructions of the program. In this case, instructions are just ordinary instructions of the CPU. These include adding and moving data. However, multiple cores can ofen run multiple instructions simultaneously. This serves to raise the overall speed of programs amenable to parallel computing. What often happens is that their manufacturers integrate cores into a unit of integrated circuit die referred to as a chip multiprocessor or even into a multiple number of dies organized into a single package (Lemstra, Hayes, Groenewegen, 2003).
Initial processors were developed with just a single core. However, multiple core processors composed of a number of cores that are larger than traditional techniques do not stand the test of efficiency any more. This is caused by the issue of congestion that results from having to supply instructions and data to several processors. As a matter of fact, the threshold for many cores lies approximately within a range of several cores, beyond which the technology of network on chip becomes disadvantageous. Conversely, a dual core processor is composed of two cores, while a quad core processor has four cores. Essentially, a multiple core processor can implement the task of multiprocessing in just one unit of physical package. In this respect, designers may choose to couple several cores into a tight or loose multicore device. For instance, different sets of cores may share or not share caches. Thus, they may implement the technology of shared memory types of inter core methods of communication. Currently, network topologies for common interconnect cores include a bus, ring, three-dimensional mesh and others. Homogenous types of multicore systems are composed only of identical cores, while heterogeneous systems have dissimilar cores (Murthy, Guruswamy).
Recent improvements in a multicore processor have to do with software algorithms and their better implementation. Specifically, gains that can possibly be achieved are restricted by the fraction of the used software in case it can be parallelized to enable them to run on multiple cores at the same time. This improvement has been described by the Amdahl’s law and may achieve speed-up factors that approach the number of cores or more, in case the problem is sufficiently split to perfectly fit within the cache of every core and technically evades the use of slow system memories. However, most applications cannot be accelerated to a large extent, unless programmers inject into a program a prohibitive amount of efforts towards factoring the whole problem. The parallelization is considered as a grey area that has a great potential for further improvements of processor technology (Cheung, Nim, Kiyoshi, Winzer, Gerhard, 1990).
Various forms of agile methodologies are based on the same philosophy, characteristics and practices. However, from the standpoint of their actual implementation, each of them is entitled to its own recipe of practices, tactics and terminologies. For instance, the scrum is described as a lightweight type of framework of management that has been used for managing and controlling all types of incremental projects. The scrum has recently gained some significant amount of popularity especially in the software community. This has been caused by the fact that it is simple, evidently productive and has a peculiar ability to be used as a wrapper for various practices in the world of engineering through other agile methodologies (Lemstra, Hayes, & Groenewegen, 2003).
There is also “Extreme Programming” that has recently entered into the market as one of the most popular agile methods. This is a principled approach meant for delivering high quality software faster and with a certain amount of continuity. It has been credited with promoting high customer involvements and ensuring rapid feedback loops. Further, there is a crystal methodology that has emerged as considerably lightweight and an adaptable approach to the idea of developing software. Essentially, the crystal methodology is itself a conglomeration of such smaller methodologies as Crystal Clear, Crystal Orange and Crystal Yellow. A striking feature of these smaller methodologies is the fact that they are driven by factors that include the size of a team, the critical nature of a system and priorities as determined in the project. Other methodologies that are quite noticeable are a dynamic system development method that has grown out of the need for setting standards for an industrial project delivery framework. Moreover, there is a feature-driven development and lean software methodologies. However, a methodology that would suit a mobile device is the Extreme Programming (XP), because its major goal is to develop for mobile devices a multimedia streaming application (Walker, Chair, 2009).
The study has come up with five online storage services or backup systems. Indeed, it is everyone’s desire to access files from any computer anywhere in the world as this is considered to be very convenient. This technology has been around for some time, but the recent entry of new generation services has made it look as a basic requirement considering it so cheap and easy to use. As a matter of fact, most of these online service providers enable their customers to share their files with preferred friends and colleagues, who may not be close to them geographically. The research has come up with several storage backup systems, which include Box.net that integrates Gmail and Zoho. Box.net can store all required documents and serve as a hub of one’s virtual office. “Live Mesh” also originates from the study findings as the only component of the latest ventures of Microsoft that is into cloud computing. Among other backup systems there are Dropbox, JungleDisk and Oosah (Lemstra, Hayes, & Groenewegen, 2003).
The research has found out the potential significance of remote backup systems that can be employed for safer data storages. This service provides its users with a workable system for storage of computer files and with a backup system. The future of technology will be shaped to a great extent by this application built on the basis of the client’s software program, which typically runs on a predetermined schedule. It collects data, compresses them and moves them into off-site hardware of a remote backup service provider. It is area field that should be investigated on further, so that its utilization can fully harness the unused potential that seems to remain untapped ("Federal Standard 1037C: Glossary of Telecommunications Terms").