The Dell PowerEdge Me Modular Blade Enclosure is built from the ground up to combat data-center sprawl and. IT complexity, delivering one of the most. An ultra-efficient blade server with up to 44 cores of Intel processing power and 24 DIMMs of the PowerEdge Me blade enclosure and PowerEdge VRTX. PE Me Blade Enclosure (includes 1xCMC and 9x12V High Efficiency Fans) Modular Integration - Blade Chassis with Factory Installed Switches.
|Language:||English, Spanish, Arabic|
|ePub File Size:||25.38 MB|
|PDF File Size:||16.47 MB|
|Distribution:||Free* [*Register to download]|
PowerEdge Me. Technical Guide. The Me chassis provides flexibility, power and thermal efficiency with scalability for future needs. and IT complexity, the. PowerEdge Me enclosure delivers one of the most energy-efficient, flexible and manageable blade server products on the market. The Dell PowerEdge Me Modular Blade Enclosure is built from the ground up to combat data center sprawl and IT complexity, delivering one of the most.
It is also possible to connect a virtual KVM switch to have access to the main-console of each installed server. The blade servers, although following the traditional naming strategy e. The blades differ in firmware and mezzanine connectors. Midplane Indication on back of chassis to see which midplane was installed in factory The midplane is completely passive. The server-blades are inserted in the front side of the enclosure while all other components can be reached via the back. The enhanced midplane 1.
The process for the system took more than three times longer than that for the. PowerEdge Blade System BladeSystem c-class Task executed Removing items from shipping boxes Installing the server in the rack Powering on the system Total time required Figure 2: Time to unpack, install, and power on both systems.
Times are in hours: minutes: seconds. Shorter times are better. We discuss this process in the following stages: Receiving the system Removing system components from their packaging o Opening and removing parts from shipping boxes o Removing components and server from outer shipping box o Removing plastic wrap and foam pieces from server o Removing chassis from pallet Installing the system in the server rack o Installing rails and mounting chassis into server rack o Installing blades into chassis o Installing power distribution unit PDU and cables Powering on the system For each stage and sub-stage after we received the systems, we enumerate the steps we took, note the amount of time each step took, and provide representative photographs.
Receiving the system As Figure 3A shows, the shipment consisted of two boxes: a large box strapped to a pallet and a smaller box.
The shipping service delivered both boxes to our second-floor lab. As Figure 3B shows, the shipment consisted of 78 boxes. Upon arrival at our building, the large box was strapped to a pallet and the 77 smaller boxes were attached to a large pallet with plastic wrap. Because the large pallet could not fit into the elevator to our second-floor office, we had to unwrap the pallet in the lobby and make multiple elevator trips to bring the smaller boxes up to our lab. Note: We did not time the process of transporting the 77 smaller boxes.
Figure 3A: The boxes upon delivery in our lab.
Figure 3B: The boxes after we brought them up to our lab. Opened and started reading instructions. Cut off the straps, then cut through the tape to open the top of the box. Opened the hard disk boxes and placed them on the table. Opened the blade server boxes and placed the blade servers on the table. Opened the blade kit boxes and placed the processors and heat sinks on the table. Opened the Ethernet switch box and placed the switch on the table. Opened the fan boxes and placed the fans on the table.
Opened the PDU boxes and placed items on the table. Figure 4B: Opening the server box. Removing components and server from outer shipping box 1.
Removed the topmost box and set it aside. Removed the outermost cardboard covering. Removed the topmost boxes and the template for inserting the enclosure onto the rack and set them aside. Removed the foam padding from the top of the server. Figure 5B: The server with outer cardboard covering removed.
Removing plastic wrap and foam pieces from server 1.
Removed the second box and set it aside. Removed the foam pieces from the blade system. Removed the corner pieces and set them aside. Removed the plastic wrap around the server.
The Drive 0, blade 1 drive was partially ejected. Figure 6B: The server with plastic wrap removed. Removing chassis from pallet 1. Firmware 4.
The external interfaces are mainly meant to be used as uplinks or stacking-interfaces but can also be used to connect non-blade servers to the network. The internal ports towards the blades are by default set as edge or "portfast" ports. Another feature is to use link-dependency. You can, for example, configure the switch that all internal ports to the blades are shut down when the switch gets isolated because it loses its uplink to the rest of the network.
All PCM switches can be configured as pure layer-2 switches or they can be configured to do all routing: both routing between the configured VLAN's as external routing. When using the switch as routing switch you need to configure vlan interfaces and assign an IP address to that vlan interface: it is not possible to assign an IP address directly to a physical interface.
To stack the new PC-Mk switch the switches need to run firmware version 4. Stacks can contain multiple switches within one Me chassis but you can also stack switches from different chassis to form one logical switch. Up to six MXL blade-switch can be stacked into one logical switch. The MXL switches also support Fibre Channel over Ethernet so that server-blades with a converged network adapter Mezzanine card can be used for both data as storage using a Fibre Channel storage system.
You can assign up to 16 x 10Gb uplinks to your distribution or core layer. Cisco offers a range of switches for blade-systems from the main vendors.
The Catalyst doesn't offer stacking virtual blade switching  Catalyst The series switches offer 16 internal 1Gb interfaces towards the blade-servers.
You can stack up to 8 Catalyst switches to behave like one single switch. This can simplify the management of the switches and simplify the spanning tree topology as the combined switches are just one switch for spanning tree considerations.
It also allows the network manager to aggregate uplinks from physically different switch-units into one logical link. The release of the B22Dell is approx. These switches would need to be connected to a "native" FCoE switch such as the Powerconnect B-series e same as a Brocade switch or a Cisco Nexus series switch with fibre channel interfaces and licenses.
It uses either the B or C fabrics to connect the Fibre Channel mezzanine card in the blades to the FC based storage infrastructure. The M offers 16 internal ports connecting to the FC Mezzanine cards in the blade-servers and 8 external ports. From factory only the first two external ports 17 and 18 are licensed: additional connections would required extra Dynamic Ports On Demand or DPOD licenses.
When delivered 12 of the ports are licensed to be operation and with additional licenses you can enable all 24 ports.
As with all other non-Ethernet based switches it can only be installed in the B or C fabric of the Me enclosre as the A fabric connects to the "on motherboard" NICs of the blades and they only come as Ethernet NICs or converged Ethernet.
The G offers 24 ports: 16 internal and 8 external ports. Unlike the M switches where the external ports are using QSFP ports for fiber transceivers, the has CX4 copper cable interfaces.
For example: if only a few of the blade-servers do use fibre-channel storage you don't need a fully manageble FC switch: you just want to be able to connect the 'internal' FC interface of the blade directly to your existing FC infrastructure.
A pass-through module has only very limited management capabilities. Managing enclosure An Me enclosure offers several ways for management. One would normally connect the Ethernet links on the CMC avoiding a switch in the enclosure. Often a physically isolated LAN is created for management allowing management access to all enclosures even when the entire infrastructure is down. The document describes various possible configurations when installing VMware Infrastructure 3.
November Dell Inc 2 Contents 1. Advantages and disadvantages for each configuration along with the steps to install are detailed. Configuration Options 2.
Blade servers can be used in a stand-alone mode Figure 1 , or they can be connected to a shared SAN through a fibre channel switch. Direct attached Fibre Channel is not supported. Provides maximum Fibre Channel bandwidth and the ability to seamlessly utilize existing Fibre Channel investments Requires additional external switches, rack space and cabling The different qualified SAN configurations are shown in Figures 2, 3 and 4. Pass-Through modules provide one connection for each blade server that is connected to external Fibre Channel switches.
Figure 3 illustrates similar configuration but using redundant Fibre Channel Switch Modules available for Dell blade server chassis.
Figure 4 shows the configuration when external Fibre Channel switches are used in addition to Fibre Channel switch modules in the chassis. Networking can be configured in various ways for traffic isolation or redundancy. A default network for Virtual Machines can be created at installation time that creates a network interface for virtual machines on NIC0.
In order to better utilize the network bandwidth and to provide redundancy, the NICs can be configured as shown in Table 2. Moderate VM Performance?