1 / 15

A Testbed for Study of Thermal and Energy Dynamics in Server Clusters

A Testbed for Study of Thermal and Energy Dynamics in Server Clusters. Shen Li, Fan Yang, Tarek Abdelzaher University of Illinois at Urbana Champaign. The Goal. Create an open shared facility for experimentation as a vehicle to promote energy management research and collaboration Focus:

gada
Download Presentation

A Testbed for Study of Thermal and Energy Dynamics in Server Clusters

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A Testbed for Study of Thermal and Energy Dynamics in Server Clusters Shen Li, Fan Yang, Tarek Abdelzaher University of Illinois at Urbana Champaign

  2. The Goal • Create an open shared facility for experimentation as a vehicle to promote energy management research and collaboration • Focus: • Investigate energy consumption on back-end servers • Investigate server resource allocation policies that minimize total energy consumption while meeting client performance demands

  3. Testbed Configuration Cluster Scale: We have 67 machines altogether, and 40 of them (most powerful machines) are made public currently. OS: CentOS5-64bit (will move to Scientific Linux 6 soon). CPU: Intel Xeon x3430, 4 core, 64bit (frequency range: 1.2GHZ ~ 2.4GHZ with 10 levels). MEM: 4GB Disk: 50GB home directory shared by all users using NFS among all servers. 200G local disk on each machine using etx3. PDU: Avocent PM 3000. CRAC: Liebert Challenger 3000. Network: 40 machines are connected as a complete graph with one switch. Thermal sensor: New PC Notebook USB Thermometer

  4. Current Capabilities * The Computer Room Air Conditioner (CRAC) is of great importance to us. Therefore, the API to manipulate CRAC set point is not public.

  5. Available Software • If you would like to minimize the interference from other uses, you can copy the binary file and install the software yourself somewhere else under /scratch. The files under /scratch will persist. • To install new software, you can use “wget” on tareka machines to download binary or source file for installation. Since you do not have sudo authority, you cannot utilize “sudo yum install”.

  6. If you have any suggestions about adding new knobs and sensors, please let us know!

  7. User Guide Step 1: have ccnx (http://www.ccnx.org/) installed and configured according to their tutorial, and make sure that your ccnd can talk to our cluster hub ndn.cs.illinois.edu Step 2: Download dcapi from here https://bitbucket.org/shenli/dcapi. Step 3: Setup dcapi according to the readme file.

  8. User Guide Step 4: Use the client GUI to check existing reservations, and find one available slot for your reservation. We have just released the first prototype of DCAPI. We don’t have any remote users yet. Above data is randomly generated for testing.

  9. User Guide Step 5: After launching the GUI, click add button and fill in the simple form to add your reservation. You can pick any user ID that matches regular expression ([a-z]|[0-9]){6,16}. It will be your CentOS user ID during your reservation.

  10. User Guide Step 6: Check your Email to get your login password.

  11. User Guide Step 2: Use ssh to login to the gateway node tareka01.cs.uiuc.edu, with your username and password. Please note that, the username and password will only be active during your reservation period. For security reasons, remote users do not have sudo access. If there is any operation you need that cannot be accomplished without sudo access, please let us know. We will add one wrapper into dcapi for that if possible.

  12. User Guide Step 2: Use dcapi as other ordinary Linux command lines. (e.g., type “dcapi get_cpu_temp”)

  13. API Design • To prevent excessive access to our sensors, dcapi uses master/worker daemons to wrap up real sensor APIs. The data is pulled every fixed time interval and cached at the daemon. • Global resources (PDU, and CRAC) are cached at master daemon. Local master daemon uses RPC to read data from master daemon.

  14. Reservation System Design

  15. For More Information • Please visit: • http://green-datacenters.web.cs.illinois.edu • More questions? • Please contact: Tarek Abdelzaher zaher@cs.uiuc.edu

More Related