LCS Value Proposition November 2014 Data Center Trends 2014 • Improving Data Center Power Efficiency & Reducing Carbon Footprint • • • • Initiatives for raising data center temperatures Data center upgrades for cooling and power management Low power server alternatives (ARM) – HP Moonshot, etc. DCIM offerings for efficient management of DC infrastructure • Increasing power density & core counts for HPC clusters with focus on energy efficiencies. • “Green 500” configurations with pervasive adoption of GPGPU and Intel Phi cores for compute • Moves to higher voltage power inputs to reduce copper and power distribution challenges • Adoption of alternatives to traditional air cooling schemes • • • Cold Plates utilizing water and ethylene water glycol fluids Heat Pipes used in conjunction with water based heat exchangers Single phase and phase change submersion cooling employing dielectric fluids • Market growth for modular approaches to add Data Center capacity • Attractive economics, flexibility, and lead time compared to brick and mortar options • Modular Data Center Market projected by 451 Research to be $2.5 Billion annually by 2015 © LCS 2014 - Proprietary and Confidential 2 LCS Cooling Complements Trends • Improving Data Center Power Efficiency • • • • Reductions in “Power to Cool” at the device level of up to 98% (Traditional Air vs LCS) Typical reductions in overall data center power consumption of 40%* Inherent ability to recover and utilize waste heat for facility purposes Potential for achieving “True PUE’s” as low as 1.02 • Increasing Power Density • Vastly superior physics for heat transfer – volumetric heat capacity ~ 1500x that of air • More compact packaging options – no need for fans • Cooling system components can be remotely configured to increase IT density • Modular Data Center Configurations • • • • Lower cost infrastructure for fluid distribution and heat exchange More compact footprints than with typical air-cooled options Lower operating costs and maintenance requirements than air cooled configurations No considerations or provisions required for local air quality – sealed systems * Based on comparison to Data Center Study of 2012 ASHRAE Report Findings © LCS 2014 - Proprietary and Confidential 3 Fans are a Problem Fans Waste Energy • 15% of total datacenter energy is used to move air • Additionally fans in the chassis can use up to 20% of IT power at the device level • Fans are inefficient and generate heat that must be removed Fans Waste Space • Racks need room to breathe • CRAC units require space around the racks Fans Reduce Reliability • • • • Fans fail Thermal fluctuations drive solder joint failures Structure borne vibration frets electrical contacts Fans expose electronics to air • Oxidation/corrosion of electrical contacts • Exposure to electrostatic discharge events • Sensitivity to ambient particulate, humidity and temperature conditions © LCS 2014 - Proprietary and Confidential 4 Benefits from Eliminating Fans Cool High Power Electronics - Lower operating temperatures result in lower leakage current - All internal components are kept within normal operating temperature ranges Save Energy - For LCS, device power-to-cool can be reduced by up to 98% vs. air-cooled devices - For LCS, the “true” cooling PUE is ~1.03 - Waste heat easily can be recovered for other uses Save Space - Higher rack density because there is no need for hot aisles or to circulate air in the racks For LCS, 64 IT devices in a 42U rack or 72 devices in a 48U rack Increase Reliability - Sealed fluid circuits prevent failures from corrosion and contamination - Liquid submersion reduces thermal fatigue of solder joints Operate Silently - Fan noise is eliminated Five Technologies Eliminate Fans 1. 2. 3. 4. 5. Clustered Systems Green Revolution Cooling Iceotope SGI LiquidCool Solutions © LCS 2014 - Proprietary and Confidential 5 Clustered Systems • Clustered Systems circulates a refrigerant through an external cold plate mounted on top of each server in the rack. • Mechanical pressure is used to create contact between the cold plate and the top of the server. • Heat is transferred to from the server to the refrigerant by conduction. • The system requires a refrigerant temperature less than 30oC. © LCS 2014 - Proprietary and Confidential 6 Green Revolution Cooling • Green Revolution Cooling’s CarnotJetTM system circulates a mineral oil based dielectric fluid through a tank containing submerged IT devices. • The system resembles a rack tipped over on its back, with modified servers inserted vertically into slots in the tank. • Each 42U tank is filled with roughly 250 gallons of the mineral oil. • Maintenance access is through the top of the tank © LCS 2014 - Proprietary and Confidential 7 Iceotope • Iceotope mounts off-the-shelf motherboards inside sealed hot-swappable cartridges that are flooded with a dielectric fluid, 3M Novec. • Novec, which remains a liquid, moves heat to a hot plate mounted on the side of the cartridge. • Water circulates in a secondary circuit through the hot plate, and heat is transferred from Novec to water in the hot plate by conduction. • Aster Capital, an investment group backed by Alstom, Schneider Electric and Solvay, recently invested $10 million in Iceotope • Along with the investment Schneider announced that it intends to commercialize Iceotope technology © LCS 2014 - Proprietary and Confidential 8 SGI • SGI uses a two-phase (evaporative) immersion cooled system. • Electronic components are submerged into a bath of Novec, an expensive dielectric refrigerant. • Boiling occurs on the surface of the heat generating devices and vapor passively rises to the top of the enclosure, where it condenses on water-cooled coils and falls back into the tank. © LCS 2014 - Proprietary and Confidential 9 LiquidCool Solutions Patented Directed-Flow Technology • No fans or other moving parts in the chassis • Total liquid submersion in a eco-friendly dielectric fluid • Rack-mounted devices are easy to maintain • Within a device “cool” liquid is circulated directly to the components with the highest power density • The remaining components are cooled by bulk flow as the dielectric fluid is drawn through the unit to a return manifold • Electronics are decoupled from the environment How it Works © LCS 2014 - Proprietary and Confidential 10 Heat Dissipation LCS Cooling System Elements • Pump supplies “cool” dielectric liquid to multiple IT racks • If there is no energy recycling option “hot” fluid is circulated to an evaporative fluid cooler • Incoming “cool” fluid can be as warm as 45°C for most applications © LCS 2014 - Proprietary and Confidential 11 Reliability LCS technology decouples electronics from the room, eliminating the root causes of failure: • Dramatic reduction in thermal fluctuations, which drive solder joint failures • Much lower operating temperatures for the board and components (typically 20-30 C cooler device temps than with air) • No oxidation/corrosion of electrical contacts • No fretting corrosion of electrical contacts induced by structural vibration • No moving parts within the device enclosure (fan failures are one of the highest service triggers for electronics) • No exposure to electrostatic discharge events • No sensitivity to ambient particulate, humidity, or temperature conditions When maintenance is required to upgrade components: • The swap out procedure takes less than 2 minutes with no measurable loss of fluid • An IT device can be removed from a rack, drained, opened, serviced, reassembled, refilled, and reinstalled within a 15 minute turnaround window © LCS 2014 - Proprietary and Confidential 12 Robust IP Portfolio 20 Issued & 19 Pending Patents • • • • • • • Liquid tight server case with dielectric liquid for cooling electronics Extruded server case used with liquid coolant of electronics Computer with fluid inlets and outlets for direct-contact liquid cooling Case and rack system for liquid submersion cooling of an array of electronic components Computer case for liquid submersion cooled components Liquid submersion cooled computer with directed flow Gravity assisted directed liquid cooling of electronics © LCS 2014 - Proprietary and Confidential 13 Additional LCS Benefits Any Shape or Size Clustered Systems, Green Revolution, SGI, Iceotope No Water Iceotope Scalable Green Revolution, SGI Easy to Swap Devices Green Revolution, SGI Easy to Maintain Devices Green Revolution, SGI, Iceotope Harsh Environment Deployments Green Revolution, SGI Costs Less than Air Clustered Systems. SGI, Iceotope Liquid Submerged Computer Operating Underwater in an Aquarium © LCS 2014 - Proprietary and Confidential 14 Any Shape or Size 8 servers with liquid-to-liquid cooling distribution unit Industrial and embedded computing applications © LCS 2014 - Proprietary and Confidential 15 Any Shape or Size Liquid Submerged Computer with Passive Radiator Liquid Submerged Computer with Stacked Boards © LCS 2014 - Proprietary and Confidential 16 64-Server Configuration Connection to remote CDU © LCS 2014 - Proprietary and Confidential 17 Low Cost “Clamshell ”Server • • • Motherboard sandwiched between two sealed enclosures Rack-mountable I/O connectors remain outside the liquid enclosure © LCS 2014 - Proprietary and Confidential 18 Clamshell Server System © LCS 2014 - Proprietary and Confidential 19 Example –Modular Data Center Air-Cooled Input Power – 500kW IT Power – 250kW LiquidCool Input Power – 320kW IT Power – 250kW © LCS 2014 - Proprietary and Confidential 20 1MW Hybrid HPC Module Overall dimensions 12’ x 12’ x 42’ © LCS 2014 - Proprietary and Confidential 21 200 KW Hybrid HPC Module Overall dimensions: 12’ x 12’ x 20’ • Air cooled section for data storage & switches Liquid cooled section for compute © LCS 2014 – Proprietary and Confidential 22 Fluid Distribution for LSS 220 Rack © LCS 2014 – Proprietary and Confidential 23 250 kW Liquid-to-Liquid CDU Fully Redundant Heat Exchangers, Pumps, and Pump Control Units Approximate size: 48” wide x 48” deep x 30” high © LCS 2014 – Proprietary and Confidential 24 Federal Data Center Upgrade Current Facility Layout (Air-Cooled) • 50’ x 100’ Facility Space • 115 racks of air cooled IT equipment • Estimated IT compute power consumption of 500 kW • 7 Air Handling Units fully operational at capacity Facility Space Reduction using LCS Servers • 50’ x 100’ Facility Space • 24 Liquid Cooled 48U racks w/72 servers/rack • 500+ kW of IT compute capacity at 21 kW/rack • Approximate footprint = 450 ft2 • Air Handling Units to Red may be decommissioned • Remainder of space available for repurposing or expansion © LCS 2014 - Proprietary and Confidential 25 TROPEC Transformative Reductions in Operational Energy Consumption TROPEC Objective – Allow enterprise servers and communication/network equipment to continuously operate in tropical environments with no mechanical cooling • LiquidCool submitted a proposal for “Modular System for High-Efficiency Electronics Cooling at Expeditionary Base Camps” in September 2013 • A comprehensive assessment of the LiquidCool system has been completed at Lawrence Berkeley National Laboratory • US Navy (PACOM) has recommended that the LiquidCool system be moved forward to the field assessment phase • Independent testing of the TROPEC system achieved successful cooling of an HPC server and cooling unit at ambient temperature of 101°F for 24 hours achieving a true PUE of 1.019 © LCS 2014 - Proprietary and Confidential 26 LCS Value Proposition Any electronics can be cooled There is no water near electronics Rack-mounted devices are hot swappable and easy to maintain There are no moving parts or heat exchange barriers in the IT device chassis There is no boiling or condensing The dielectric liquid is inexpensive, eco-friendly and never needs to be replaced There are no microchannels to clog Electronics are not exposed to air pollution Blade and rack fans are eliminated There is no noise, vibration or extreme temperature fluctuation Air handlers and CRAC units are eliminated Mechanical refrigeration is eliminated There is no need for humidification or dehumidification Raised floors and high ceilings are eliminated Heat is recovered in a convenient form for recycling …and LCS cooling costs less too! © LCS 2014 - Proprietary and Confidential 27 More Information Herb Zien, CEO [email protected] 414-803-6010 Jay Ford, VP Sales & Marketing [email protected] 847-370-7296 Rick Tufty, VP Engineering [email protected] 507-535-5829 Steve Einhorn, Chairman [email protected] 414-453-4488 © LCS 2014 - Proprietary and Confidential 28
© Copyright 2024