Free Essay

Data

In:

Submitted By Bornslippynuxx
Words 806
Pages 4
William Wragg
Mr. Thomas Fortenberry
IST 113471
17 October 2014

Chapter 7 Exercises 1. What is the definition of a local area network? a. A communication network that interconnects a variety of data communicating devices within a small geographic area and broadcasts data at high data transfer rates with very low error rates. 2. List the primary activities and application areas of local area network. b. File server, print server, connections to other networks. 3. List the advantages and disadvantages of local area networks. c. The advantages are sharing of files and devices, and intercommunication. d. The disadvantages are maintenance, complexity, and costs. 4. What are the basic layouts of local area networks? List two advantages that each layout has over the others. e. Buses advantages are the use of low-noise coaxial cable, and inexpensive taps. f. Star-wired buses are simple to interconnect, have easy to add components, and are most popular. g. Star-wired rings are simple to interconnect and have easy to add components. 5. What is meant by a passive device? h. A signal that enters is neither amplified nor regenerated. The signal is simply passed on. 6. What is meant by a bidirectional signal? i. A signal that propagates in either direction on a medium. 7. What are the primary differences between baseband technology and broadband technology? j. Baseband is a digital signal, where broadband is analog and can carry many signals. 8. What purpose does a hub serve? k. The hub is a collection point for workstations. 9. What is the difference between a physical design and a logical design? l. Physical is the wiring and components. Logical is how the software passes the data. 10. What is a medium access control protocol? m. The software that allows a workstation to insert its data onto the LAN 11. What are the basic operating principles behind CSMA/CD? n. CSMA/CD: Listen to medium; if no one transmitting, transmit. Continue to listen for collisions. If someone is transmitting, wait. 12. What is meant by a “nondeterministic” protocol? o. You cannot determine precisely when a workstation will get a chance to transmit due to potential collisions.

13. What does the term 100BaseT stand for? p. One hundred mega-bits per second transmission over digital signals, using twisted pair wiring. 14. What is the difference between Fast Ethernet and regular Ethernet? q. Fast Ethernet transmits at 100 Mbps, while regular Ethernet transmits at 10 Mbps. 15. What are the latest 10-Gbps Ethernet standards? r. 10GBase-fiber, 10GBase-T, 10GBase-CS. 16. What is the primary advantage of power over Ethernet? The primary disadvantage? s. The primary advantage is not having to run a separate power line to power the device. t. The disadvantage is making sure the switch has enough power to run PoE devices. 17. How does a transparent switch work? u. Through a process called backwards learning the switches learn where workstations are and then use that information for future forwarding decisions. 18. What is the purpose of a virtual LAN? v. To setup a network within a network with different rules, bandwidth optimization, and network security.

19. How does a switch encapsulate a message for transmission? w. The switch looks at NIC/MAC addresses and forwards accordingly. 20. When referring to a hub or a switch, what is a port? x. The port is the connection that is used to connect a workstation or another hub or switch to this hub or switch. 21. What are the basic functions of a switch? y. A switch examines a packet’s destination address and routes the packet to the appropriate workstation. 22. How does a switch differ from a hub? z. A switch examines addresses. A hub does not. 23. What is cut-through architecture? {. The device is passing the data packet on before it has even finished entering the device. 24. How is a full-duplex switch different from a half-duplex switch? |. Transmits on two pairs of wires, and sends and receives on two different signals. 25. What is meant by link aggregation? }. The combining of two or more data paths, or links, into one higher-speed link.

Thinking outside the Box #5
After the latest round of corporate reorganizations, you now have six internal divisions instead of seven. The technology department tries to organize the network layout by keeping users in each division on their own segment of the network, but it is getting tired of rearranging the wire in the closets to accommodate these changes. What might you suggest that the technology department use instead to support the six internal divisions? a. I would suggest they implement Virtual Networks for some of the less concrete divisions and use that as a way to not have to move wiring every time to accommodate changes within the corporation.

Similar Documents

Premium Essay

Data

...Discuss the importance of data accuracy. Inaccurate data leads to inaccurate information. What can be some of the consequences of data inaccuracy? What can be done to ensure data accuracy? Data accuracy is important because inaccurate data leads may lead to such things as the closing down of business, it may also lead to the loosing of jobs, and it may also lead to the failure of a new product. To ensure that one’s data is accurate one may double check the data given to them, as well as has more than one person researching the data they are researching. Project 3C and 3D Mastering Excel: Project 3G CGS2100L - Section 856 MAN3065 - Section 846 | | 1. (Introductory) Do you think Taco Bell was treated fairly by the mass media when the allegations were made about the meat filling in its tacos? I think so being that they are serving the people for which I must say that if you are serving the people then it’s in the people rights to know what exactly you are serving them. 2. (Advanced) Do you think the law firm would have dropped its suit against Taco Bell if there were real merits to the case? It’s hard to say but do think that with real merits it would have changed the playing feel for wit real merits whose the say that Taco Bell wouldn’t have had an upper hand in the case. 3. (Advanced) Do you think many people who saw television and newspaper coverage about Taco Bell's meat filling being questionable will see the news about the lawsuit being withdrawn? I doubt that...

Words: 857 - Pages: 4

Free Essay

Data

...Import Data from CSV into Framework Manager 1. Save all your tables or .csv file under one folder. In our case we will use the Test folder saved on blackboard with three .csv files named TestData_Agent.csv, TestData_Customer.csv, TestData_InsuranceCompany.csv. 2. Now , locate the correct ODBC exe at “C:\Windows\SysWOW64\odbcad32.exe” 3. Once the ODBC Data Source Administrator is open, go to the “System DSN” tab and click “Add”. 4. Select “Microsoft Text Driver (*.txt, *.csv)” if you want to import from csv files. 5. Unclick the “Use Current Directory”, and then click Select Directory to define the path of your data source. Give data source a name as well. Let’s use TestData in this case. NOTE: All the files under the specified location will be selected by default. 6. Again press ok and close the dialogue. Now we will import this Database/csv files into Cognos using Framework Manager. 7. Now Go to find Framework Manager. C:\Program Files (x86)\ibm\Cognos Express Clients\Framework Manager\IBM Cognos Express Framework Manager\bin 8. Right click on 'FM.exe', and then select 'Properties'. Click 'Compatibility' tab. Check "Run this program as an administrator' under 'Privilege Level'.  9. Open Framework Manager and create a new project and give it any name, in this case CSV_MiniProject. Then click OK. 10. Put the username: “Administrator” and password:”win7user”. 11. Select Language as English and hit ok. 12. Select Data Sources...

Words: 775 - Pages: 4

Free Essay

Data

...instructed to backfill with temporary labour. The collated data is being used to investigate the effect of this shift in labour pattern, paying particular attention to staff retention. The table below gives a month by month record of how many staff have been employed, temporary and permanent , how many temporary staff are left at the end of each month compared to how many are left that are on a permanent contract. Month | Temporary staff | permanent staff | total | permanent leavers | Temporary leavers | total leavers | Jan-15 | 166 | 359 | 525 | 7 | 2 | 9 | Feb-15 | 181 | 344 | 525 | 15 | 5 | 20 | Mar-15 | 181 | 344 | 525 | 0 | 7 | 7 | Apr-15 | 204 | 321 | 525 | 23 | 7 | 30 | May-15 | 235 | 290 | 525 | 31 | 12 | 43 | Jun-15 | 238 | 287 | 525 | 3 | 17 | 20 | Jul-15 | 250 | 275 | 525 | 12 | 42 | 54 | Aug-15 | 267 | 258 | 525 | 17 | 23 | 40 | Sep-15 | 277 | 248 | 525 | 10 | 27 | 37 | Oct-15 | 286 | 239 | 525 | 9 | 30 | 39 | Nov-15 | 288 | 237 | 525 | 2 | 34 | 36 | Dec-15 | 304 | 221 | 525 | 16 | 45 | 61 | Jan-16 | 305 | 220 | 525 | 1 | 53 | 54 | Feb-16 | 308 | 217 | 525 | 3 | 57 | 60 | An explanation of how I analysed and interpreted the data To make a comparison between the labour pattern and retention, I placed the above data into a line graph this gives a more of an idea to trends over the period My Findings The actual level of staff has remained constant throughout the data collated, as each job requires a specific amount of man...

Words: 621 - Pages: 3

Free Essay

Data

...Data Collection - Ballard Integrated Managed Services, Inc. (BIMS) Learning Team C QNT/351 September 22, 2015 Michael Smith Data Collection - Ballard Integrated Managed Services, Inc. (BIMS) Identify types of data collected--quantitative, qualitative, or both--and how the data is collected. A survey was sent out to all the employees’ two paychecks prior and a notice to complete the survey was included with their most recent paychecks. After reviewing the surveys that have been returned it was found that the data collected is both quantitative and qualitative. Questions one thru ten are considered qualitative data because the response for those questions are numbered from one (very negative) to five (very positive), which are measurements that cannot be measured on a natural numerical scale. They can only be classified or grouped into one of the categories and are simply selected numerical codes. Then, questions A-D could fall under quantitative data because it can determine the number of employees in each department, whether they are male or female and amount of time employed with the company. From that data it is able to find an average of time employed, then subcategorize by department, gender and if they are a supervisor or manager. Identify the level of measurement for each of the variables involved in the study. For qualitative variable there are a couple levels of measurements. Questions A, C, and D in Exhibit A fall in nominal-level data because when asking...

Words: 594 - Pages: 3

Premium Essay

Big Data and Data Analytics

...Big Data and Data Analytics for Managers Q1. What is meant by Big Data? How is it characterized? Give examples of Big Data. Ans. Big data applies to information that can’t be processed or analysed using traditional processes or tools or software techniques. The data which is massive in volume and can be both structured or unstructured data. Though, it is a bit challenging for enterprises to handle such huge amount fast moving data or one which exceeds the current processing capacity, still there lies a great potential to help companies to take faster and intelligent decisions and improve operations. There are three characteristics that define big data, which are: 1. Volume 2. Velocity 3. Variety * Volume: The volume of data under analysis is large. Many factors contribute to the increase in data volume, for example, * Transaction-based data stored through the years. * Unstructured data streaming in social media. * Such data are bank data (details of the bank account holders) or data in e-commerce wherein customers data is required for a transaction. Earlier there used to data storage issues, but with big data analytics this problem has been solved. Big data stores data in clusters across machines, also helping the user on how to access and analyse that data. * Velocity: Data is streaming in at unprecedented speed and must be dealt with in a timely manner. RFID tags, sensors and smart metering are driving the need to deal with...

Words: 973 - Pages: 4

Premium Essay

Knbs Data

...KNBS DATA DISSEMINATION AND ACCESS POLICY November 2012 VISION A centre of excellence in statistics production and management MISSION To effectively manage and coordinate the entire national statistical system to enhance statistical production and utilization Herufi House, Lt. Tumbo lane P.O. Box 30266 – 00100 GPO Nairobi, Kenya Tel: +254-20-317583/86/88,317612/22/23/51 Fax: +254 – 20-315977 Email: info@knbs.or.ke Web: www.knbs.or.ke i WI-83-1-1 Preface Kenya National Bureau of Statistics (KNBS) is the principal agency of the Government for collecting, analysing and disseminating statistical data in Kenya. KNBS is the custodian of official statistical information and is mandated to coordinate all statistical activities, and the National Statistical System (NSS) in the country. Official statistics are data produced and disseminated within the scope of the Statistical Programme of the National Statistical System (NSS) in compliance with international standards. To achieve this mandate, KNBS strives to live up to the aspirations of its vision; to be a centre of excellence in statistics production and management. Chapter Four on The Bill of Rights section 35 of the new constitution in Kenya gives every citizen right of access to information held by the State. This policy document strives to provide a framework for availing statistical information to the public in conformity with this bill and government’s open data initiative. This...

Words: 3544 - Pages: 15

Free Essay

Data Information

...Data and Information Summary HCI/520 11/18/2013 Data and Information Summary Today we live in a world where data is a critical resource. Information is also a critical resource and consists of data that is processed into meaningful information for the purpose of organizations and users. Collected data is stored into what is known as databases where it is organized into potentially valuable information. Data also known as Raw data is a stream of facts that are not organized or arranged into a form that people can understand or use (Gillenson, Ponniah, Kriegel, Trukhnov, Taylor, Powell, & Miller, 2008) . Raw Data are facts that have not yet been processed to reveal their meaning (Gillenson, Ponniah, Kriegel, Trukhnov, Taylor, Powell, & Miller, 2008). For example when AT&T wireless ask their clients to participate in a survey about the products they have purchased or how was their customer service experience the data collected is useful but not until the raw data is organized by combining it with other similar data and analyzed into meaningful information. Information is the result of processing raw data to reveal its meaning (Coronel, Morris, & Rob, 2010). Data processing can be as simple as organizing data to reveal patterns or as complex as making forecasts or drawing inferences using statistical modeling (Gillenson, Ponniah, Kriegel, Trukhnov, Taylor, Powell, & Miller, 2008). Both data and information are types of knowledge which share similarities...

Words: 538 - Pages: 3

Free Essay

Bad Data

...Bad Data AMU DEFM420 Big Data refers to volume, variety and velocity of available data. The issue with that is that any emphasis is put on volume or quantity of data. The quantity is a very vague element of Big Data, there is no precise requirements to purely volume based data. What should be considered in big data, the complexity and depth of the data? If the content of data is deep and containing detailed information it holds more purpose. When we analyzes data we as a culture prefer less to review but of more importance. I would rather read two pages of relevant data then read one hundred pages that contain 3 pages of data. This is a factor of human nature but also a business factor. The majority of what we do in government work is time sensitive. We operate on a system of end dates. With time being a factor wasting time on Big Data that isn’t always pertinent is a waste. While in cases of no time limit, having the full three V’s of big data is acceptable and may in the end give more accurate information after spending excessive time sorting through the information mainly the volume portions. Is the system of Big Data wrong? No it is not wrong but the concept is too vague. For different situations data needs to be limited. Others not so much so it gives us a system and collection of information that is in some cases excessive for the need. It is a double edged sword. There are other aspects of Big Data collections useful in contracting offices...

Words: 325 - Pages: 2

Premium Essay

Data Management

...University Data Management March 18, 2014 Data partitioning is a tool that can help manage the day-to-day needs of an organization. Each organization has unique values that drive business. All organizations have policies and processes that are influenced by their environment and industry. The use of data partitioning can help productivity by recognizing the need to categorize data to tailor unique needs. This approach does require some effort. To transition to a new database approach, organizations need to assess the pros and cons of a database transition. The scale of an organization’s database may be the one factor that drives an adoption of this approach. Data partitioning has been developed to address issues that traditional database queries have created. One main problem that partitioning was created to solve is the performance of database queries. According to Alsultanny (2010), “System performance is one of the problems where a significant amount of query processing time is spent on full scans of large relational tables. Partitioning of the vertical or horizontal partitioning or a combination of both is a reliable way of solving the problem” (p.273). By separating queries into either horizontal or vertical processes, the user can avoid delays and strains on a database. This saves time which can be used to improve the productivity an organization has towards their day-to-day operations. Large-scale databases receive the most benefits from partitioning. Data partitioning...

Words: 1572 - Pages: 7

Free Essay

Data & Information

...able to function without data, information and knowledge. Data, information and knowledge are different from one another, yet there are interrelated to each other. Data Data are unprocessed raw facts which can be either qualified and or quantified. It explains phenomenal refer to statistical observations and other recordings or collections of evidence (Chaim Zins, 2007, p.480). Data can be in numbers or text. For example, temperature currency, gender, age, body weight. Figure 1, is example of data recorded in Microsoft Excel data sheet. Figure 1 Information The outcome of data processing is information. Figure 2 expresses the process of how data is being transformed to information. Data which is the input when being processed such as organized, examined, analyzed, summarized it gives the output as information. Information is processed data which gives explicit meaning to its readers. Based on Figure 1 data, after processed them, gives you the information of the percentage of a group of 24 youth, the number of times they eat fast food in a week as shown in Figure 3. Figure 3 show that youth in their twenties eats fast food at least once a week there are even a small number of them (4.1%) takes fast food almost every day (6 times/ week). It gives the information about the demand of fast food among youth in their twenties. Figure 3 The average age of this group also can be obtained from the data in the excel data sheet in Figure 4. Figure...

Words: 285 - Pages: 2

Free Essay

Big Data

...Lecture on Big Data Guest Speaker Simon Trang Research Member at DFG RTG 1703 and Chair of Information Management Göttingen University, Germany 2014 The City City of Göttingen • Founded in the Middle Ages • True geographical center of Germany • 130,000 residents Chair of Information Management Lecture on Big Data at Macquarie University 2 2 The University Georg-August-Universität Göttingen (founded in 1737) • • • • One of nine Excellence Universities in Germany 13 faculties, 180 institutes 26,300 students (2013) 11.6% students from abroad (new entrants: approximately 20%) • 13,000 employees (including hospital and medical school), including 420 professors • 115 programs of study from A as in Agricultural Science to Z as in Zoology are offered (73 bachelor / 22 master programs) Chair of Information Management Lecture on Big Data at Macquarie University 3 “The Göttingen Nobel Prize Wonder” Over 40 Nobel prize winners have lived, studied, and/or lived, studied or/and researched 41 Prize researched at the University of Göttingen, among them… at the University of Göttingen, among them… • • • • • • • • • • • • • • Max von Laue, Physics, 1914 Max von Laue, physics, 1914 Max Planck, physics, 1918 Max Planck, Physics, 1918 Werner Heisenberg, physics, 1932 Werner Heisenberg, Physics, 1932 Otto Hahn, chemistry 1944 Otto Hahn, Chemistry 1944 Max Born, physics, 1954 Max Born, Physics, 1954 Manfred Eigen, chemistry, 1967 Manfred Eigen, Chemistry, 1967 Erwin...

Words: 1847 - Pages: 8

Premium Essay

Data Gathering

...Four Basic Data Gathering Procedure Options Some students are not aware of the fact that they need to know some data gathering proceduretechniques when writing their research papers. Usually, they are simply concentrating on how to come with a good thesis statement, how to develop literature review or even how to cute reference materials. Let us now talk about the methods of data gathering since some research proposal examples do not even mention about this segment. There are different ways for you to conduct data gathering procedures. Usually, these ways are related to the same processes in statistics. Dissertation research methods are almost always related to data gathering so you really need to learn how to acquire your data for analysis. * Data mining – this procedure involves the search for published data from reputable sources. The process is simpler than other techniques but you need to make sure that the data is up to date. * Interviewing o this data gathering procedure involves a certain amount of time and effort investment. However, you can maximize the data that you can acquire form each respondents because you will personally acquire data from them. * Depending on your research paper topics, you can conduct surveying. If you wish to gather quick and raw data, this is the best medium for you. Prepare a set of questionnaire and then have your respondents fill them out. * Lab experiments – this type of data gathering procedure is intended if you wish...

Words: 267 - Pages: 2

Free Essay

Data Breaches

...Daniel Baxter Nico Ferragamo Han Vo Romilla Syed IT 110 8 December 2015 Data Breaches The Case In July of 2014 JPMorgan Chase, a multinational banking and financial services holding company was hacked. JPMorgan Chase is the largest bank in the United States, the sixth largest bank in the world, and the world’s third largest public company. Initial reports from JPMorgan Chase stated that the attack had only breached about one million accounts. Further details revealed that the hack breached the accounts of seventy-six million households (roughly two-thirds of the total number of households in the United States), and about seven million small businesses. While the hack began in July, it was not fully stopped until the middle of August, and it was not disclosed to the public until September. The hack is considered to be one of the most serious attacks on an American Corporation’s information systems and is one of the largest data breaches in history. JPMorgan Chase claims that the login information associated with the accounts (such as social security numbers and passwords) was not compromised, and the information that was stolen had not been involved in any fraudulent activities, however, the names, email addresses, physical addresses, and phone numbers on the accounts were taken by the hackers. The hack was believed to have been committed by a group of Russian hackers. It’s also believed to have been part of a large ring of attempted attacks on as many as nine banks and...

Words: 1557 - Pages: 7

Premium Essay

Primary Data vs Secondary Data

...Differences Between Primary Data vs Secondary Data -Submitted by Arvind Kartik SOURCES OF PRIMARY DATA Regardless of any difficulty one can face in collecting primary data; it is the most authentic and reliable data source. Following are some of the sources of primary data. Experiments Experiments require an artificial or natural setting in which to perform logical study to collect data. Experiments are more suitable for medicine, psychological studies, nutrition and for other scientific studies. In experiments the experimenter has to keep control over the influence of any extraneous variable on the results. Survey Survey is most commonly used method in social sciences, management, marketing and psychology to some extent. Surveys can be conducted in different methods. Questionnaire: It is the most commonly used method in survey. Questionnaires are a list of questions either an open-ended or close -ended for which the respondent give answers. Questionnaire can be conducted via telephone, mail, live in a public area, or in an institute, through electronic mail or through fax and other methods. Interview : It is a face-to-face conversation with the respondent. It is slow, expensive, and they take people away from their regular jobs, but they allow in-depth questioning and follow-up questions. The interviewer can not only record the statements the interviewee speaks but he can observe the body language or non-verbal communication such as face-pulling, fidgeting...

Words: 659 - Pages: 3

Free Essay

Big Data

...A New Era for Big Data COMP 440 1/12/13 Big Data Big Data is a type of new era that will help the competition of companies to capture and analyze huge volumes of data. Big data can come in many forms. For example, the data can be transactions for online stores. Online buying has been a big hit over the last few years, and people have begun to find it easier to buy their resources. When the tractions go through, the company is collecting logs of data to help the company increase their marketing production line. These logs help predict buying patterns, age of the buyer, and when to have a product go on sale. According to Martin Courtney, “there are three V;s of big data which are: high volume, high variety, high velocity and high veracity. There are other sites that use big volumes of data as well. Social networking sites such as Facebook, Twitter, and Youtube are among the few. There are many sites that you can share objects to various sources. On Facebook we can post audio, video, and photos to share amongst our friends. To get the best out of these sites, the companies are always doing some type of updating to keep users wanting to use their network to interact with their friends or community. Data is changing all the time. Developers for these companies and other software have to come up with new ways of how to support new hardware to adapt. With all the data in the world, there is a better chance to help make decision making better. More and more information...

Words: 474 - Pages: 2