Skip to main content

Green Datacenter

Note: This blog is written with help of my friend Rajanikanth Katikitala


Green IT
•       Use servers with less power demands. The lower the PUE, the more efficient data center infrastructure
Power Usage Efficiency =       Total facility Power 
        -------------------------------            
      Total IT equipment Power
•       Floor space reduction by moving to high density blade servers
•       Use hot & cold aisle containments to avoid mixing of hot air with cold air
•       Use rack  or row cooling instead of traditional CRAC units
•       Use free cooling. Retrofit outside air supply where possible
•       Right size the cooling requirement as per the IT load to avoid over sizing cooling.
•       use high density area to focus on cooling
•       Distribute high-density racks throughout the layout to mitigate hot
•       spots; use spot rack cooling as necessary




The Uptime Institute “tiers”
 The uptime Institute classifies data centers according into “tiers”.  The following table is taken from the paper, “Industry Tier Classifications Define Site Infrastructure Performance,” by W. Pitt Turner IV and Kenneth Brill, both of the UpTime Institute.

Tier I
Tier II
Tier III
Tier IV
Number of delivery paths (of power and a/c etc)
Only 1
Only 1
1 active
1 passive
2 active
Redundant components
N
N+1
N+1
2(N+1) or S+S
Support space to raised floor ratio
20%
30%
80-90%
100%
Initial watts/sq ft
20-30
40-50
40-60
50-80
Ultimate watts/sq ft
20-30
40-50
40-60
150+
Raised floor height
12”
18”
30-36”
30-36”
Floor loading pounds/sq ft
85
100
150
150+
Utility voltage
208, 480
208, 480
12-15Kv
12-15Kv
Months to implement
3
3-6
15-20
15-20
Year first deployed
1965
1970
1985
1995
Construction $/sq ft raised floor
$450
$600
$900
$1,100
Annual IT downtime due to site
28.8 hrs
22.0 hrs
1.6 hrs
0.4 hrs
Site availability
99.671%
99.749%
99.982%
99.995%
“Even a fault-tolerant and concurrently maintainable Tier IV site will not satisfy an IT requirement of “Five Nines” (99.999%) uptime. The best a Tier IV site can deliver over time is 99.995%, and this assumes a site outage occurs only as a result of a fire alarm or EPO (Emergency Power Off), and that such an event occurs no more than once every five years. Only the top 10 percent of Tier IV sites will achieve this level of performance.  Unless human activity issues are continually and rigorously addressed, at least one additional failure is likely over five years.  While the site outage is assumed to be instantaneously restored (which requires 24 x “forever” staffing), it can still require up to four hours for IT to recover information availability.
Tier IV’s 99.995% uptime is an average over five years.  An alternative calculation using the same underlying data is 100% uptime for four years and 99.954% for the year in which the downtime event occurs.  Higher levels of site uptime can be achieved by protecting against accidental activation or the real need for fire protection and EPOs. Preventatives include high sensitivity smoke detection, limiting fire load, signage, extensive training, staff certification, limiting the number of “outsiders” in critical spaces, and treating people well to increase pride in their work. All of these measures, if taken, can reduce the risk of failures.  Other solutions include placing the redundant parts of the IT computing infrastructure in different site infrastructure compartments so that a site infrastructure event cannot simultaneously affect all IT systems.  Another alternative is focusing special effort on business-critical and mission-critical applications so they do not require four hours to restore.  These operational issues can improve the availability offered by any data center, and are particularly important in a “Four Nines” Tier IV data center housing IT equipment requiring “Five Nines” availability.”
Heating Ventilation and Air Conditioning standards
The Uptime Institution recommends hot and cold aisles.  Instead of having the fronts of all racks facing one way, racks in raised floor data centers should be installed with the fronts towards each other (the cold aisle) and the backs towards each other (the hot aisle). 
ASHRAE "TG9 HDEC" is the industry guiding reference document for data center environments in the US. According to it the data center ambient temperature should be from 20o C (68o F) to 25o C (77o F) and measured in the cold aisle at 5 ft. above the raised floor at the air intake of the operating equipment.  The relative humidity should be from 40 to 55%.  The future TIA-942 standard will be comparable to the ASHARE standard. The TIA-942 standard will call for a maximum rate of change of 5o C (9o F) per hour.
 “Operationally, an errpr on the slightly cooler (21-22 deg C 70-72deg F) room ambient measured at an average 3-5' off the floor in the center of the aisles.  Experience has shown that prolonged temperatures in the room above 75-78 deg. F (24-26C) reduce HDD life up to 50% and cause serious internal issues as interior processor temperatures can exceed 40 deg. C (105 deg. F)”.
If you are involved in writing a proposal that calls for data center temperatures, it is strongly recommended that SLAs state that the HVAC system will keep equipment within the temperature ranges specified by OEM equipment manufacturers, rather than the air temperatures mentioned above.  If a loss of city water supply causes a cooling emergency, you will bless that bit of advice.
Rack spacing
Based on studies that Jonathon Jew performed at 12 web hosting data centers between 30% to 60% of the computer room space can be consumed by power & cooling equipment and circulation (compare with the UpTime table’s entries for Support space to raised floor ratio).  He comments that: “… a lot depended on the heat density of the computer room, the shape of the computer room, and the quantity and size of columns.” 
That reference to circulation concerns rack spacing and other passageways. The ASHRAE and ANSI/TIA recommendations for cabinet/rack spacing are to provide a minimum of 4 feet clearance in the cold aisles in front of cabinets and 3 feet clearance in the hot aisles behind cabinets.  The Americans with Disabilities Act dictates a minimum of 3 feet for any aisles where wheel chairs need to traverse.  Jonathon says: “For cabinets/racks with equipment with a depth no greater than 2 feet, you may be able to get by with rows of cabinets/racks spaced on 5 feet centers.  However, the recommended spacing based on ASHRAE and ANSI/TIA is 7 ft centers.”
If your customer has HP 42 U racks (24” wide and close to 40” deep) with 4 foot of clearance for the cold aisle and the 3 foot of clearance in the hot aisle there will be two racks occupying 164” x 24”.  That’s 13.7 square feet per rack including the hot and cold aisles but ignoring any other additional passageways

Popular posts from this blog

Data Center Migration

Note: This blog is written with the help of my friend Rajanikanth
Data Center Migrations / Data Center Consolidations
Data Center Consolidations, Migrations are complex projects which impact entire orgnization they support. They usually dont happen daily but once in a decade or two. It is imperative to plan carefully, leverage technology improvements, virtualization, optimizations.
The single most important factor for any migration project is to have high caliber, high performing, experienced technical team in place. You are migrating business applications from one data center to another and there is no scope for failure or broken application during migration. So testing startegy should be in place for enterprise business applications to be migrated.
Typical DCC and Migrations business objectives
Business Drivers
·Improve utilization of IT assets ·DC space & power peaked out - business growth impacted ·Improve service levels and responsiveness to new applications ·Reduce support complexi…

HP CSA Implementation

I know the above picture is little confusing but don’t worry I break it down and explain in detail. By the time I am done explaining you all will be happy. HARDWARE AND SOFTWARE REQUIREMENTS 1.VMware vSphere infrastructure / Microsoft Hyper V: For the sake of Simplicity we will use VMware vSphere. We Need vSphere 4.0 /5/5.5 and above and vCenter 4.0 and above ready and installed. This is the first step. 2.We need Software medias for HP Cloud Service Automation, 2.00, HP Server Automation, 9.02, HP Operations Orchestration (OO)9.00.04, HP Universal CMDB 9.00.02, HP Software Site Scope, 11.01,HP Insight Software6.2 Update 1 3.DNS, DHCP and NTP systems are already installed and configured. NTP information should be part of VM templates 4.SQL Server 2005 or Microsoft® SQL Server 2008 or Microsoft® SQL Server 2012 , Oracle 11g, both 32-bit and 64-bit versions may be used for CSA database.
5.We will install  HP Cloud Service Automation, 2.00, HP Server Automation, 9.02, HP Operations Orchestra…

Openstack- Its importance in Cloud. The HP Helion Boost

Every enterprise expects few things from cloud computing, mainly:

· Auto scaling: The workload should increase and decrease as needed by the IT environment.

· Automatic repair: If there is any fault or crash of the application or the server, it automatically fix it

· Fault tolerant: The application or underlying technology is intelligent enough to make itself fault torrent

· Integrated lifecycle: It should have integrated lifecycle

· Unified management: Its easy to manage all different aspects of technology

· Less cost

· Speed


Its year 2014. till now only 5% to 7% enterprises are using cloud computing. Such a small number. Its a huge opportunity and a vast majority for anyone who is interested in providing cloud computing services.
Current IT environment is very complex. You just cant solve all your problems with cloud computing.
There are legacy systems, databases, data processors, different hardware and software. You name it , there are so many technology available in just o…