Storage system hardware overview
The VSP G1x00 and VSP F1500 storage systems can be configured with one or two controller chassis and up to 12 drive chassis that contain the drives. A controller chassis contains the control logic, processors, memory, and interfaces to the drive chassis and the host servers. A drive chassis contains drives, power supplies, and the interface circuitry that connects it to the controller.
The controller rack contains the controller chassis and up to two drive chassis. Additional racks can contain an intermix of 16U SFF or LFF drive chassis and/or 8U flash drive chassis as well as one or two Hitachi NAS (HNAS) file system servers. When the controllers of a two-controller system are housed in separate racks, the two controller racks can be placed up to 100 meters apart. In addition, the drive racks attached to a controller rack can be placed up to 100 meters from the controller rack.
The VSP G1x00 storage systems support the latest hard disk drives (HDDs), solid-state drives (SSDs), and flash module drives (FMDs). The drives are installed into a specific drive chassis according to the drive type. The VSP G1x00 also support a diskless configuration (without drive chassis, all external storage).
The VSP F1500 all-flash array is equipped with advanced high-density FMDs that provide up to 4M IOPS and 40 PB of capacity for multi-workload consolidation.
Storage system configurations and model list
The following section describes the storage system configurations, models, and components.
The illustration shown is only an example. The storage system provides flexibility for placing the controller and drive chassis within the racks. For more information about system configurations, contact your sales account representative.
Item |
Description |
1 |
10U controller chassis |
2 |
(Optional) maximum of twelve 8U, FMD chassis |
3 |
16U SFF or LFF drive chassis |
4 |
8U space |
Chassis |
Description |
Maximum chassis / drives per system | |
Two-controller system |
Single-controller system | ||
Controller |
10U x 1 controller |
1 |
2 |
SFF |
16U x 192 2.5-inch HDDs |
6 / 1,152 |
12 / 2,304 |
LFF |
16U x 96 3.5-inch HDDs |
6 / 576 |
12 / 1152 |
FMD |
8U x 48 FMDs |
6 / 288 |
12 / 576 |
SFF / LFF |
SSDs |
1921 1,1522 |
3841 2,3042 |
1Maximum number in a standard performance configuration. 2Maximum number in a high-performance configuration. |
Number of VSD pairs |
1 |
2 or 3 |
4 |
5 |
6 or 7 |
8 |
Number of controller chassis |
1 |
1 |
1 |
2 |
2 |
2 |
Minimum cache memory (GB) |
64 |
64 |
64 |
64 |
128 |
128 |
Cache backup kit |
1 |
1 |
1 |
1 |
2 |
2 |
Front-end directors |
1 |
1 |
1 |
1 |
1 |
1 |
Back-end directors |
0 |
0 |
0 |
0 |
0 |
0 |
Number of racks |
1 |
1 |
1 |
11 |
11 |
11 |
Drive chassis (SFF/LFF/FMD) |
Optional |
Optional |
Optional |
Optional |
Optional |
Optional |
1Assumes both controllers are housed in a single rack. Two separate racks are required if the controllers are separately housed. |
The list provides the parts description number for each storage system model.
Component |
Model number | |||
VSP G1000 |
VSP G1500 (upgrade from VSP G1000 to G1500) |
VSP G1500 |
VSP F1500 | |
Primary Controller Chassis | DKC-810I-CBXA/ DKC-810I-CBXAC | DKC-810I-CBXA/ DKC-810I-CBXAC |
DKC-810I-CBXE/DKC810I-CBXET |
DKC-810I-CBXE/DKC810I-CBXET |
Second Controller Chassis | DKC-810I-CBXB | DKC810I-CBXB |
DKC-810I-CBXF |
DKC-810I-CBXF |
Controller Chassis Bezel | DKC-F810I-BCH | DKC-F810I-BCH | DKC-F810I-BCH | DKC-F810I-BCH |
Badge for bezel | 5557156-001 | N/A | 5562191-001 | 5562488-001 |
DKC Power Cord Kit (USA) | DKC-F810I-PLUC | DKC-F810I-PLUC | DKC-F810I-PLUC | DKC-F810I-PLUC |
DKC Power Cord Kit (EU) | DKC-F810I-PLEC | DKC-F810I-PLEC | DKC-F810I-PLEC | DKC-F810I-PLEC |
DKC Power Cord Kit (China) | DKC-F810I-PLCC | DKC-F810I-PLCC | DKC-F810I-PLCC | DKC-F810I-PLCC |
LFF Drive Chassis | DKC-F810I-UBX/ DKC-F810I-UBXC | DKC-F810I-UBX/ DKC-F810I-UBXC | DKC-F810I-UBXC | N/A |
SFF Drive Chassis | DKC-F810I-SBX/ DKC-F810I-SBXC | DKC-F810I-SBX/ DKC-F810I-SBXC | DKC-F810I-SBXC | DKC-F810I-SBXC |
Drive Chassis Bezel | DKC-F810I-BUH | DKC-F810I-BUH | DKC-F810I-BUH | DKC-F810I-BUH |
DKU Power Cord Kit (USA) | DKC-F810I-PHUC | DKC-F810I-PHUC | DKC-F810I-PHUC | DKC-F810I-PHUC |
DKU Power Cord Kit (EU) | DKC-F810I-PHEC | DKC-F810I-PHEC | DKC-F810I-PHEC | DKC-F810I-PHEC |
DKU Power Cord Kit (China) | DKC-F810I-PHCC | DKC-F810I-PHCC | DKC-F810I-PHCC | DKC-F810I-PHCC |
FMD Chassis | DKC-F810I-FBX | DKC-F810I-FBX | DKC-F810I-FBX | DKC-F810I-FBX |
FMD Chassis Bezel | DKC-F810I-BFH | DKC-F810I-BFH | DKC-F810I-BFH | DKC-F810I-BFH |
FBX Power Cord Kit (USA) | DKC-F810I-PFUC | DKC-F810I-PFUC | DKC-F810I-PFUC | DKC-F810I-PFUC |
FBX Power Cord Kit (EU) | DKC-F810I-PFEC | DKC-F810I-PFEC | DKC-F810I-PFEC | DKC-F810I-PFEC |
FBX Power Cord Kit (China) | DKC-F810I-PFCC | DKC-F810I-PFCC | DKC-F810I-PFCC | DKC-F810I-PFCC |
Additional Service Processor | DKC-F810I-SVP/ DKC-F810I-SVPC | DKC-F810I-SVP/ DKC-F810I-SVPC | DKC-F810I-SVPC | DKC-F810I-SVPC |
DKC-F810I-SVP101/DKC-F810I-SVP10T | DKC-F810I-SVP101/DKC-F810I-SVP10T | DKC-F810I-SVP101/DKC-F810I-SVP10T | DKC-F810I-SVP101/DKC-F810I-SVP10T | |
Additional Hub | DKC-F810I-HUB | DKC-F810I-HUB | DKC-F810I-HUB | DKC-F810I-HUB |
Intercontroller Connecting Kit | DKC-F810I-MOD5 | DKC-F810I-MOD5 | DKC-F810I-MOD5 | DKC-F810I-MOD5 |
Intercontroller Connecting Cable | DKC-F810I-MFC5 | DKC-F810I-MFC5 | DKC-F810I-MFC5 | DKC-F810I-MFC5 |
Intercontroller Connecting Kit | DKC-F810I-MOD30 | DKC-F810I-MOD30 | DKC-F810I-MOD30 | DKC-F810I-MOD30 |
Intercontroller Connecting Cable | DKC-F810I-MFC30 | DKC-F810I-MFC30 | DKC-F810I-MFC30 | DKC-F810I-MFC30 |
Intercontroller Connecting Kit | DKC-F810I-MOD1J | DKC-F810I-MOD1J | DKC-F810I-MOD1J | DKC-F810I-MOD1J |
Intercontroller Connecting Cable | DKC-F810I-MFC1J | DKC-F810I-MFC1J | DKC-F810I-MFC1J | DKC-F810I-MFC1J |
Device Interface Cable CC1 | DKC-F810I-CC1 | DKC-F810I-CC1 | DKC-F810I-CC1 | DKC-F810I-CC1 |
Device Interface Cable CC2 | DKC-F810I-CC2 | DKC-F810I-CC2 | DKC-F810I-CC2 | DKC-F810I-CC2 |
Device Interface Cable CC4 | DKC-F810I-CC4 | DKC-F810I-CC4 | DKC-F810I-CC4 | DKC-F810I-CC4 |
Device Interface Cable FC5 | DKC-F810I-FC5 | DKC-F810I-FC5 | DKC-F810I-FC5 | DKC-F810I-FC5 |
Device Interface Cable FC30 | DKC-F810I-FC30 | DKC-F810I-FC30 | DKC-F810I-FC30 | DKC-F810I-FC30 |
Device Interface Cable FC1J | DKC-F810I-FC1J | DKC-F810I-FC1J | DKC-F810I-FC1J | DKC-F810I-FC1J |
iSCSI 8-port 10G Host Adapter | DKC-F810I-8IS10 | DKC-F810I-8IS10 | DKC-F810I-8IS10 | DKC-F810I-8IS10 |
Fibre Channel 16-port 8G Host Adapter | DKC-F810I-16FC8 | DKC-F810I-16FC8 | DKC-F810I-16FC8 | DKC-F810I-16FC8 |
Fibre Channel 8-port 16G Host Adapter | DKC-F810I-8FC16 | DKC-F810I-8FC16 | DKC-F810I-8FC16 | DKC-F810I-8FC16 |
Fibre Channel 16-port 16G Host Adapter | DKC-F810I-16FC16 | DKC-F810I-16FC16 | DKC-F810I-16FC16 | DKC-F810I-16FC16 |
Mainframe Fibre 16-port 8G Host Adapter for Shortwave | DKC-F810I-16MS8 | DKC-F810I-16MS8 | DKC-F810I-16MS8 | DKC-F810I-16MS8 |
Mainframe Fibre 16-port 8G Host Adapter for Longwave | DKC-F810I-16ML8 | DKC-F810I-16ML8 | DKC-F810I-16ML8 | DKC-F810I-16ML8 |
Mainframe Fibre 16-port 16G Host Adapter for Shortwave | DKC-F810I-16MS16 | DKC-F810I-16MS16 | DKC-F810I-16MS16 | DKC-F810I-16MS16 |
Mainframe Fibre 16-port 16G Host Adapter for Longwave | DKC-F810I-16ML16 | DKC-F810I-16ML16 | DKC-F810I-16ML16 | DKC-F810I-16ML16 |
FCOE 16-port Host Adapter | DKC-F810I-16FE10 | DKC-F810I-16FE10 | DKC-F810I-16FE10 | DKC-F810I-16FE10 |
Additional Cache Path Control Adapter | DKC-F810I-CPEX | DKC-F810I-CPEX | DKC-F810I-CPEX | DKC-F810I-CPEX |
Cache Memory (16 GB) | DKC-F810I-CM16G | DKC-F810I-CM16G | DKC-F810I-CM16G | DKC-F810I-CM16G |
Cache Memory (32 GB) | DKC-F810I-CM32G | DKC-F810I-CM32G | DKC-F810I-CM32G | DKC-F810I-CM32G |
Cache Backup Module Kit for Small Memory | DKC-F810I-BKMS | DKC-F810I-BKMS | DKC-F810I-BKMS | DKC-F810I-BKMS |
Cache Backup Module Kit for Large Memory | DKC-F810I-BKML | DKC-F810I-BKML | DKC-F810I-BKML | DKC-F810I-BKML |
Cache Flash Memory (128 GB) | DKC-F810I-BMM128 | DKC-F810I-BMM128 | DKC-F810I-BMM128 | DKC-F810I-BMM128 |
Cache Flash Memory (256 GB) | DKC-F810I-BMM256 | DKC-F810I-BMM256 | DKC-F810I-BMM256 | DKC-F810I-BMM256 |
Cache Flash Memory (256 GB) | DKC-F810I-BMZ256 | DKC-F810I-BMZ256 | DKC-F810I-BMZ256 | DKC-F810I-BMZ256 |
Disk Adapter | DKC-F810I-SCA | DKC-F810I-SCA | DKC-F810I-SCA | DKC-F810I-SCA |
Encryption Disk Adapter | DKC-F810I-ESCA | DKC-F810I-ESCA | DKC-F810I-ESCA | DKC-F810I-ESCA |
FIPS140-2 Level 2 Upgrade Toolkit | DKC-F810I-FIPS2 | DKC-F810I-FIPS2 | DKC-F810I-FIPS2 | DKC-F810I-FIPS2 |
Additional Processor Blades | DKC-F810I-MP | N/A | N/A | N/A |
MP2 Upgrade Kit | N/A | DKC-F810I-MP2UGH | N/A | N/A |
Additional Processor Blades 2 | N/A | DKC-F810I-MP2 | DKC-F810I-MP2 | DKC-F810I-MP2 |
Hard disk drives | ||||
300-GB SFF disk drive 15k | DKC-F810I-300KCM | DKC-F810I-300KCM | DKC-F810I-300KCMC | N/A |
DKC-F810I-300KCMC | DKC-F810I-300KCMC | |||
600-GB SFF disk drive 15k | DKC-F810I-600KGM | DKC-F810I-600KGM | DKC-F810I-600KGM | N/A |
600-GB SFF disk drive 10k | DKC-F810I-600JCM | DKC-F810I-600JCM | DKC-F810I-600JCMC | N/A |
DKC-F810I-600JCMC | DKC-F810I-600JCMC | |||
600-GB LFF disk drive 10k | DKC-F810I-600J5M | DKC-F810I-600J5M | DKC-F810I-600J5MC | N/A |
DKC-F810I-600J5MC | DKC-F810I-600J5MC | |||
900-GB SFF disk drive 10k | DKC-F810I-900JCM | DKC-F810I-900JCM | DKC-F810I-900JCMC | N/A |
DKC-F810I-900JCMC | DKC-F810I-900JCMC | |||
1.2-TB SFF disk drive 10k | DKC-F810I-1R2JCM | DKC-F810I-1R2JCM | DKC-F810I-1R2JCMC | N/A |
DKC-F810I-1R2JCMC | DKC-F810I-1R2JCMC | |||
1.8-TB SFF disk drive 10k | DKC-F810I-1R8JGM | DKC-F810I-1R8JGM | DKC-F810I-1R8JGM | N/A |
2.4-TB SFF disk drive 10K | N/A | DKC-F810I-2R4JGM | DKC-F810I-2R4JGM | N/A |
4-TB LFF disk drive 7.2k | DKC-F810I-4R0H3M | DKC-F810I-4R0H3M | DKC-F810I-4R0H3MC | N/A |
DKC-F810I-4R0H3MC | DKC-F810I-4R0H3MC | |||
6-TB LFF disk drive 7.2k | DKC-F810I-6R0H9M | DKC-F810I-6R0H9M | DKC-F810I-6R0H9M | N/A |
Solid-state drives | ||||
400-GB MLC SFF SSD | DKC-F810I-400MCM | DKC-F810I-400MCM | DKC-F810I-400MCM | N/A |
400-GB MLC LFF SSD | DKC-F810I-400M5M | DKC-F810I-400M5M | DKC-F810I-400M5M | N/A |
800-GB MLC SFF SSD | DKC-F810I-800MCM | DKC-F810I-800MCM | DKC-F810I-800MCM | N/A |
960-GB MLC SFF SSD | DKC-F810I-960MGM | DKC-F810I-960MGM | DKC-F810I-960MGM | DKC-F810I-960MGM |
1.9-TB MLC SFF SSD | DKC-F810I-1R9MGM | DKC-F810I-1R9MGM | DKC-F810I-1R9MGM | DKC-F810I-1R9MGM |
3.8-TB MLC SFF SSD | DKC-F810I-3R8MGM | DKC-F810I-3R8MGM | DKC-F810I-3R8MGM | DKC-F810I-3R8MGM |
7.6-TB TLC SFF SSD | N/A | DKC-F810I-7R6MGM | DKC-F810I-7R6MGM | DKC-F810I-7R6MGM |
15-TB TLC SFF SSD | N/A | DKC-F810I-15RMGM | DKC-F810I-15RMGM | DKC-F810I-15RMGM |
Flash module drives (FMD) | ||||
1.75-TB FMD | DKC-F810I-1R6FM | DKC-F810I-1R6FM | N/A | N/A |
DKC-F810I-1R6FN | DKC-F810I-1R6FN | DKC-F810I-1R6FN | DKC-F810I-1R6FN | |
3.5-TB FMD | DKC-F810I-3R2FM | DKC-F810I-3R2FM | N/A | N/A |
DKC-F810I-3R2FN | DKC-F810I-3R2FN | DKC-F810I-3R2FN | DKC-F810I-3R2FN | |
7-TB FMD HD | DKC-F810I-6R4FN | DKC-F810I-6R4FN | DKC-F810I-6R4FN | DKC-F810I-6R4FN |
DKC-F810I-7R0FP | DKC-F810I-7R0FP | DKC-F810I-7R0FP | DKC-F810I-7R0FP | |
14-TB FMD HD | DKC-F810I-14RFP | DKC-F810I-14RFP | DKC-F810I-14RFP | DKC-F810I-14RFP |
7-TB FMD-HDE | N/A | DKC-F810I-7R0FPE2 | DKC-F810I-7R0FPE2 | DKC-F810I-7R0FPE2 |
14-TB FMD-HDE | N/A | DKC-F810I-14RFPE2 | DKC-F810I-14RFPE2 | DKC-F810I-14RFPE2 |
Notes: 1Microsoft Windows 10 IoT Enterprise is installed. 2Supports embedded FMD accelerated compression capabilities with data-at-rest encryption. |
System controller chassis
The following figures show the front and rear views of a system controller chassis, and the following tables list the description of the components.
Item |
Name |
Min |
Max |
Description |
1 |
Control Panel |
1 |
1 |
See Power control panel. |
2 |
Cache Backup Module (BKM) |
2 |
4 |
Backup memory modules are installed in pairs and referred to as a backup memory kit. Each module contains two batteries and either a 128 GB SSD (small kit) or a 256 GB SSD (large kit). If the power fails, the cache is protected from data loss by the backup batteries and the cache flash memory (SSD). The batteries keep the cache active up to 32 minutes while the data is copied to the SSD. |
3 |
|
2 (1 pair) |
8 (4 pairs) |
A VSD can contain either an Intel Xeon 2.1 GHz or 2.3 GHz 8-core microprocessor. The VSDs are independent of the front-end directors and back-end directors, and can be shared across them. The VSDs must be installed in pairs and the VSDs control the front-end directors, back-end directors, PCI Express interface, local memory, and communication to the SVP. |
4 |
Cooling fan (intake) |
5 |
5 |
The five intake fans on the front of the controller pull air into the controller and distribute it across the controller components. |
5 |
Cache Path Control Adapter (CPA) |
1 |
4 |
The CPA uses the built-in switch to connect the VSDs to the front-end directors, back-end directors, and the cache backup memory. It distributes data (data routing function) and sends hot-line signals to the VSD. The shared memory is located on the first CPA cache board in each cluster in the primary controller. |
Item |
Name |
Min |
Max |
Description |
1 |
Power supply |
2 |
4 |
200-240 VAC input. Provides power to the controller chassis in a redundant configuration. Each power supply contains two cooling fans to ensure constant cooling if one fan fails. |
2 |
Service processor (SVP) |
1 |
2 |
A custom PC monitoring and controlling the storage system. It contains the Device Manager - Storage Navigator software, which configures and monitors the system. Connecting the SVP to a service center enables the storage system to be remotely monitored and maintained by the support team. |
3 |
Service processor (SVP) or Hub |
0 |
1 |
This space can be empty or can contain either a second SVP or a hub. If a second SVP is installed, the primary SVP is the active SVP, and the secondary SVP is the hot idle SVP with active Windows. A hub facilitates the transfer of information from the VSD pairs to the primary SVP. |
4 |
Back-end director or (optional) front-end director |
0 if diskless |
4 |
Connects the HDDs, SSDs, and FMDs. It controls the data transfer between the drives and the cache. VSP G1000, VSP G1500, and VSP F1500 support two types of back-end directors:
Important: Each back-end director and front-end director consists of a set of two of blades. See Flexible front-end director installation for details. |
2 with drives | ||||
5 |
Front-end director (host I/O module) |
1 |
2 to 5 with drives |
A front-end director (FED) provide ports that support connectivity to the open and mainframe systems belonging to the customer. In addition, some of the FEDs support virtualization of externally attached storage, remote replication between VSP G1000, VSP G1500, and VSP F1500 and other storage systems, including communication between two Hitachi storage systems in a global active device cluster. See Front-end directors for details. |
6 if diskless system | ||||
6 |
Cooling fan (exhaust) |
5 |
5 |
The exhaust fans on the rear of the controller pull hot air away from the components and push it out the back of the rack. |
7 |
FED or VSD slot |
0 |
4 |
The slots on the controller support both front-end directors and VSDs. Both FEDs and VSDs must be installed in pairs. |
Note: 1. Achieved FIPS 140-2 Level 1 certification. |
Front-end directors
A front-end director (FED) is a pair of blades installed in the controller.
The FED connects the storage system to the host servers, processes channel commands from hosts, manages host access to the cache, and controls the transfer of data between the hosts and the controller cache.
The following FEDs are available:
- iSCSI
- Fibre Channel
- FICON (shortwave and longwave)
- Fibre Channel over Ethernet (FCoE)
The Fibre Channel FED can be configured with either shortwave or longwave host connectors. The FICON is configured with either longwave or shortwave connectors that match the wavelength of the mainframe ports.
The following figure shows the port LEDs of a FED, and the following table lists the description of the port LEDs.
Item |
Name |
Color |
Description |
1 |
Blade Status |
Dark (off) Red (on) |
OFF: Power is not supplied to the system. The system is not operational. ON: Board failure. The blade can be replaced while the system is running. |
2 |
Power supply Status |
Dark (off) Amber (on) |
OFF: Power is not supplied to the system or, if power is supplied to the system, power supply in this blade is operational. ON: Power supply failure, abnormal voltage in power supply. |
3 |
Port Status (FC/iSCSI) |
Dark (off) Green (on) |
OFF: If system power is off, the port is not ready. OFF: If system power is on, the port is ready. ON: Link is active. |
4 |
Link Activity (FC/iSCSI) |
Dark (off) Green (on) |
OFF: No link activity, for three possible reasons: power is off, initialization is not completed, and if system is operational, the port is not being accessed. ON (steady): Link is available and initialization is complete, but connection to the host has not been established. Blinking: When the port is being accessed and data is being transferred between the host and the cache. |
3 |
Port Status (FICON) |
Dark (off) Green (on) |
OFF: If system power is off, the port is not ready. ON: Link is available and initialization is complete, but connection to the host has not been established. ON: Link is active. |
4 |
Link Activity (FICON) |
Dark (off) Amber (on) |
OFF: No link activity because either power is off, initialization is not complete, or, if the system is operational, the port is not being accessed. ON (fast blink): When the port is being accessed and data is being transferred between the host and the cache. |
Supported connectors and protocols
A variety of FED options are available for installation in the controller chassis. The maximum number of ports configurable in a two controller system by FED type are as follows:
- 96 iSCSI ports (10 Gbps, 8-port)
- 192 Fibre Channel ports (16 Gbps, 16-port)
- 192 Fibre Channel ports (8 Gbps, 16-port)
- 96 Fibre Channel ports (16 Gbps, 8-port)
- 176 FICON ports (16 Gbps, 16-port) available in longwave and shortwave versions
- 176 FICON ports (8 Gbps, 16-port) available in longwave and shortwave versions
- 192 FCoE ports (10 Gbps, 16-port)
See Site preparation for information about port configurations.
The following figure shows the SFP transceiver of a host connector.
Item |
Description |
1 - Host connector lock lever |
Black lever = shortwave Blue lever = longwave |
Fibre Channel, iSCSI and Fibre Channel over Ethernet (FCoE) FEDs support open system hosts while FICON FEDs support mainframe systems.
The following tables lists the supported FEDs and protocols.
FED |
System |
Ports |
Mode |
Description |
8-port, 10-Gbps iSCSI |
Open |
8 (4 per blade) |
Target and Initiator1 |
Contains two 4-port front-end modules. Each port contains a short wavelength (multi-mode) host connector. |
16-port, 16-Gbps Fibre Channel, shortwave and longwave adapter |
Open |
16 (8 per blade) |
Target and Initiator1 |
Contains two 8-port front-end modules. Each port contains an LC-type short wavelength, multimode host connector. These are installed by default. A long wavelength host connector (-1PL8) for single mode can be used instead of the short wavelength host connector. |
16-port, 8-Gbps Fibre Channel, shortwave and longwave adapter |
Open |
16 (8 per blade) |
Target and Initiator1 |
Contains two 8-port front-end modules. Each port contains an LC-type short wavelength, multi-mode host connector. These are installed by default. A long wavelength host connector (-1PL8) for single mode may be used instead of the short wavelength host connector. |
8-port, 16-Gbps Fibre Channel, shortwave and longwave adapter |
Open |
8 (4 per blade) |
Target and Initiator1 |
Contains two 4-port front-end modules. Each port contains an LC-type short wavelength, multi-mode host connector. These are installed by default. A long wavelength host connector (-1PL16) for single mode may be used instead of the short wavelength host connector. |
16-port, 8-Gbps FICON shortwave adapter |
Mainframe |
16 (8 per blade) | Target |
Contains two 8-port front-end modules. Each port contains a short wavelength (multi-mode) host connector. |
16-port, 8-Gbps FICON longwave adapter |
Mainframe |
16 (8 per blade) | Target |
Contains two 8-port front-end modules. Each port contains an LC-type long wavelength (multi-mode) host connector. |
16-port, 16-Gbps FICON shortwave adapter |
Mainframe |
16 (8 per blade) | Target |
Contains two 8-port front-end modules. Each port contains a short wavelength (multi-mode) host connector. |
16-port, 16-Gbps FICON longwave adapter |
Mainframe |
16 (8 per blade) | Target |
Contains two 8-port front-end modules. Each port contains an LC-type long wavelength (multi-mode) host connector. |
16-port, 10-Gbps Fibre Channel over Ethernet (FCoE) |
Open |
16 (8 per blade) |
Target and Initiator1 |
Contains two 8-port front-end modules. Each port contains a short wavelength (multi-mode) host connector. |
Notes:
|
Flexible front-end director installation
The maximum number of FEDs that can be installed in the controller depends on the number of VSD pairs and FEDs (BED) that are also installed.
As shown in the following figure, the controller is built with 12 I/O slots, eight of which are dual-purpose. This provides a wide range of configuration flexibility.
The following list describes the purpose of the slots.
- Eight of the 12 I/O slots support different types of devices.
- Slots 1C/2C and 1D/2D are reserved for the first two FEDs.
- Slots 1E/2E and 1F/2F support the installation of either FEDs or VSDs.
- Slots 1A/2A and 1B/2B support the installation of either BEDs or VSDs.
The following table shows the order of FED installation. If the storage system includes internal drives, the controller requires a minimum of a single pair of BEDs and can be configured to support up to two BED pairs. A storage system that does not include any internal drives is referred to as a diskless configuration. The term standard describes a controller configured with a single BED, while high performance describes a controller configured with two BED pairs.
Installation slot location | ||||||
Installation order (see the previous figure) |
Diskless model |
Standard model |
High-performance model | |||
DKC810I- CBXA/CBXAC |
DKC-F810I- CBXB |
DKC810I- CBXA/CBXAC |
DKC-F810I- CBXB |
DKC810I- CBXA/CBXAC |
DKC-F810I- CBXB | |
1 | 1PC/2PC (CHA0) | 1PJ/2PJ (CHA6) | 1PC/2PC (CHA0) | 1PJ/2PJ (CHA6) | 1PC/2PC (CHA0) | 1PJ/2PJ (CHA6) |
2 | 1PD/2PD (CHA1) | 1PK/2PK (CHA7) | 1PD/2PD (CHA1) | 1PK/2PK (CHA7) | 1PD/2PD (CHA1) | 1PK/2PK (CHA7) |
3 | 1PE/2PE (CHA2/MPB3) | 1PL/2PL (CHA8/MPB7) | 1PE/2PE (CHA2/MPB3) | 1PL/2PL (CHA8/MPB7) | 1PE/2PE (CHA2/MPB3) | 1PL/2PL (CHA8/MPB7) |
4 | 1PF/2PF (CHA3/MPB2) | 1PM/2PM (CHA9/MPB6) | 1PF/2PF (CHA3/MPB2) | 1PM/2PM (CHA9/MPB6) | 1PF/2PF (CHA3/MPB2) | 1PM/2PM (CHA9/MPB6) |
5 | 1PB/2PB (DKA1/CHA4) | 1PH/2PH (DKA3/CHA10) | 1PB/2PB (DKA1/CHA4) | 1PH/2PH (DKA3/CHA10) | — | — |
6 | 1PA/2PA (DKA0/CHA5) | 1PG/2PG (DKA2/CHA11) | — | 1PG/2PG (DKA2/CHA11) | — | — |
Installation slot location | ||||||
Installation order (see the previous figure) |
Diskless model |
Standard model |
High-performance model | |||
DKC810I-CBXE |
DKC810I-CBXF |
DKC810I-CBXE |
DKC810I-CBXF |
DKC810I-CBXE |
DKC810I-CBXF | |
1 | 1PC/2PC (CHA0) | 1PJ/2PJ (CHA6) | 1PC/2PC (CHA0) | 1PJ/2PJ (CHA6) | 1PC/2PC (CHA0) | 1PJ/2PJ (CHA6) |
2 | 1PD/2PD (CHA1) | 1PK/2PK (CHA7) | 1PD/2PD (CHA1) | 1PK/2PK (CHA7) | 1PD/2PD (CHA1) | 1PK/2PK (CHA7) |
3 | 1PE/2PE (CHA2/MPB3) | 1PL/2PL (CHA8/MPB7) | 1PE/2PE (CHA2/MPB3) | 1PL/2PL (CHA8/MPB7) | 1PE/2PE (CHA2/MPB3) | 1PL/2PL (CHA8/MPB7) |
4 | 1PF/2PF (CHA3/MPB2) | 1PM/2PM (CHA9/MPB6) | 1PF/2PF (CHA3/MPB2) | 1PM/2PM (CHA9/MPB6) | 1PF/2PF (CHA3/MPB2) | 1PM/2PM (CHA9/MPB6) |
5 | 1PB/2PB (DKA1/CHA4) | 1PH/2PH (DKA3/CHA10) | 1PB/2PB (DKA1/CHA4) | 1PH/2PH (DKA3/CHA10) | — | — |
6 | 1PA/2PA (DKA0/CHA5) | 1PG/2PG (DKA2/CHA11) | — | 1PG/2PG (DKA2/CHA11) | — | — |
Supported speeds and cable lengths
The Fibre Channel FEDs can be configured with either shortwave or longwave host connectors that match the cables connecting it to the host systems.
The following table lists the Fibre Channel cable length requirements for the front-end directors for the storage system.
Data transfer rate |
OM2 cable (50/125 μm multi-mode fiber) |
OM3 cable (50/125 μm laser optimized, multimode fiber) |
OM4 cable (50/125 μm laser optimized, multimode fiber) |
MBps |
feet / meters |
feet / meters |
feet / meters |
200 |
984.3 / 300 |
1640.4 / 500 |
N/A |
400 |
492.1 / 150 |
1246.7 / 380 |
1312.4 / 400 |
800 |
164 / 50 |
492.1 / 150 |
623.4 / 190 |
1600 |
118 / 35 |
328 / 100 |
410.1 / 125 |
Data transfer rate ( MBps) |
OM3 cable length (km) |
200, 400, 800, 1600 |
10 |
Back-end director
A back-end director (BED) is a pair of blades installed into the controller chassis and it controls the data transfer between the cache memory and internal drives of the storage system.
Hitachi offers the following BEDs:
- Standard back-end director
- Encrypting back-end director
The standard back-end director blades are each equipped with four 6-Gbps SAS ports and do not support encryption.
The encrypting back-end director (EBED) blades each provide four 6-Gbps SAS ports. When writing data to the internal drives of the system, the EBED encrypts the data. The encrypted data-at-rest is unencrypted by the EBED as it is read from the drive. The EBED is certified as FIPS 140-2 Level 2 compliant to meet the strict security standards of customers managing storage systems. A Hitachi Encryption License Key must be installed during installation in order to enable the encryption functionality of the EBED and an additional license key known as FIPS 140-2 Level 2 License Key must be installed for operating in compliance with the FIPS 140-2 Level 2 specification. For more information about the encrypting back-end directors and implementing a storage system with FIPS 140-2 Level 2 compliance, contact a Hitachi Vantara representative.
For more information about FIPS 140-2 criteria and certificate for VSP G1x00, see the following websites:
- FIPS 140-2: http://csrc.nist.gov/groups/STM/cmvp/standards.html
- FIPS 140-2 Level 2 certificate #2727 for the VSP G1x00: http://csrc.nist.gov/groups/STM/cmvp/documents/140-1/1401val2016.htm#2727
The hardware components used in the standard back-end director blades are different from the encrypting back-end director blades. A BED pair may not consist of one standard and one encrypting blade.
Flexible back-end director installation
Use the following guideline when installing standard and encrypting back-end directors (BEDs).
A maximum of two BEDs can be installed in both the primary controller chassis and secondary controller chassis for a total of four BEDs per system. Only one type of BEDs can be installed into a system. The standard and encrypting BEDs cannot be mixed in the storage system. The maximum number of BED pairs that can be installed into a controller chassis will depend on the number of FEDs that will also be installed into the controller chassis.
A system without internal drives (also known as a diskless configuration) does not require the installation BEDs. However, the system must include one or more FED pairs to provide ports to connect to externally attached storage as well as to host systems.
For more information about configuration rules for FEDs and BEDs, see Flexible front-end director installation.
Drive chassis
The VSP G1x00 support three different drive chassis. The VSP F1500 only supports the FMD chassis. All components in the drive chassis are configured with redundant pairs to prevent system failure. While the storage system is in operation, all components in the drive chassis can be added or replaced. For detailed information about the drives in each chassis, see Storage system specifications.
Drive chassis |
Description | Drive trays / Drives per tray |
Maximum number of chassis / drives per system | |
Single controller (3 racks) |
Dual controller (6 racks) | |||
SFF |
A 16U group of eight 2U drive trays. Each holds up to 24 vertically positioned 2.5-inch HDD and SSD drives, for a total of 192 SFF drives per chassis. See Figure 1: SAS connection diagram of Rack-00 (SFF/LFF standard model). | Eight 2U trays, up to 24 drives each |
6 / 1,152 hard drives Up to 192 SSDs1 Up to 1,152 SSDs2 |
12 / 2,304 drives Up to 384 SSDs1 Up to 2,304 SSDs2 |
LFF |
A 16U group of eight 2U drive trays. Each holds up to 12 horizontally positioned 3.5-inch drives, for a total of 96 LFF drives per chassis. See Figure 1: SAS connection diagram of Rack-00 (SFF/LFF standard model). | Eight 2U trays, up to 12 drives each |
6 / 576 hard drives Up to 192 SSDs1 Up to 1,152 SSDs2 |
12 / 1,152 Up to 384 SSDs1 Up to 2,304 SSDs2 |
FMD |
An 8U group of four 2U drive trays. Each holds up to 12 horizontally positioned drives, for a total of 48 FMDs per chassis. See Figure 3: SAS connection diagram of Rack-00 (FBX standard model). | Four 2U trays, up to 12 drives each |
6 / 288 |
12 / 576 |
1Maximum number in a standard-performance configuration. 2Maximum number in a high-performance configuration. |
The following illustrations show the front and rear panels of the three types of 2U drive trays, and the following tables describe the connectors and LEDs.
Item |
Name |
Color |
Description |
1 |
POWER LED |
Green |
OFF: Power is not supplied to the system. ON: Power is supplied to the system. |
READY LED |
Green |
OFF: System is not operational. ON: Normal operation. Storage system is operational. Fast blink - internal processing. Storage system is operational. Slow blink - offline download processing completed (maintenance). | |
LOCATE LED |
Orange |
OFF: Normal operation ON: Nonfatal error. Storage system can remain operating. Contact technical support. See Getting help in the preface of this manual. | |
2 |
ALM LED (alarm) |
Red |
OFF: Normal operation ON: Fatal error. Contact technical support. See Getting help in the preface of this manual. |
3 |
ACT LED (Active) |
Green |
OFF: Drive is not being accessed. Blinking: Drive is being accessed. |
Item |
Name |
Color |
Description |
1 |
POWER LED |
Green |
OFF: No power is supplied to the system. ON: Power is supplied to the system. |
READY LED |
Green |
ON: Normal operation. Storage system is operational. Fast blink: Internal processing. Storage system is operational. Slow blink: Offline download processing completed (maintenance). | |
LOCATE LED |
Orange |
ON: Nonfatal error. Storage system can continue operating. Contact technical support. See Getting Help in the preface of this manual. | |
2 |
ENC IN LED |
Green |
ON: Port is connected to an OUT port in the controller. This can be directly or via another drive box with daisy-chained cables. |
3 |
ENC IN connector |
- |
Connects the drives to the ENC OUT port in the control chassis either directly or via another drive box with daisy-chained cables. |
4 |
ENC OUT connector |
- |
Connects the drives to the ENC IN port in the control chassis either directly, or via another drive box with daisy-chained cables. |
5 |
ENC OUT LED |
Green |
ON: Indicates that the port is connected to an IN port in the controller. This can be performed directly or indirectly, as previously described. |
6 |
Console port |
- |
RJ-45 connector (not used) |
7 |
Power Supply |
- |
Converts 200 VAC to the DC voltages used by the drives and the ENC adapters. |
8 |
RDY (Ready) LED |
Green |
OFF: No power is supplied to the system or the power supply has failed. ON: The power supply is operating normally. |
9 |
AC IN LED |
Green |
ON: AC input is normal. |
10 |
ALM (Alarm) LED |
Red |
Power supply has failed. Contact technical support. See Getting Help in the preface of this manual. |
11 |
AC Socket |
- |
For IEC60320-C14 plug: 200 - 240 VAC +8% - 6% 50/60 Hz |
Item |
Description |
1 |
Flash module Active LED - Lights when the flash module is activated - Blinks at drive access. |
2 |
Flash module Alarm LED - Lights when the flash module has an error and should be replaced. |
3 |
SAS / ENC Module Power LED |
4 |
SAS / ENC Module Alarm LED - Indicates fatal error condition. |
5 |
SAS / ENC standard IN connector |
6 |
SAS / ENC high performance IN connector |
7 |
ENC adapter - Connects the flash modules to the BEDs in the controller through ENC cables. |
8 |
SAS / ENC standard OUT connector |
9 |
SAS / ENC high performance OUT connector |
10 |
Power cord receptacle |
11 |
Power Supply - 220 VAC input, draws approximately 265 watts. |
12 |
Power Supply Ready 1 LED - Lights when 12 VDC power #1 is ready. |
13 |
Power Supply Ready 2 LED - Lights when 12 VDC power #2 is ready. |
14 |
Power Supply alarm LED - Lights when power supply has an error. |
Cache memory
The VSP G1000, VSP G1500, and VSP F1500 storage systems can be configured with 64 GB to 1 TB of cache memory per controller. The cache memory is installed in one or two cache path control adapters (CPA). A CPA feature consists of a pair of redundant blades that are installed and work together to provide cache and shared memory for the system. The following figure shows two CPAs (2-3, and 1-4).
Cache memory modules (DIMMs) are available in either 16 GB or 32 GB sizes. The minimum memory required per controller is 64 GB, either two 16 GB DIMMs or one 32 GB DIMM must be installed in each CPA blade. The memory modules in a system must all be the same size.
The following table shows minimum and maximum cache capacities per controller. The figures are doubled for a two-controller system.
# of controller chassis configuration |
Capacity of cache memory module |
Maximum cache capacity configurable based on the number of cache path control adapter pairs included: | |
1 cache path control adapter pair (included with controller) |
2 cache path control adapter pairs (one included with controller and an additional feature added) | ||
VSP G1000, VSP G1500, and VSP F1500 including a Primary controller chassis only |
16 GB |
256 GB |
512 GB |
32 GB |
512 GB |
1024 GB | |
VSP G1000, VSP G1500, and VSP F1500 including both a Primary and Secondary controller chassis |
16 GB |
512 GB |
1024 GB |
32 GB |
1024 GB |
2048 GB | |
Notes:
|
Item |
Description |
Item |
Description |
2 and 3 |
Main (required) cache path control adapters |
1 and 4 |
Optional cache path control adapters |
1 and 2 |
Cluster 1 |
3 and 4 |
Cluster 0 |
Memory operation
The controller places all read and write data into the cache. The amount of fast-write data in cache is dynamically managed by the cache control algorithms to provide an optimum amount of read and write cache, depending on the workload read and write I/O characteristics.
Data protection
The VSP G1000, VSP G1500, and VSP F1500 storage systems protect the loss of data or configuration information stored in the cache when electrical power fails. The cache is kept active for up to 32 minutes by the cache backup batteries while the system configuration and data are copied to the cache flash memory in the cache backup modules. For more information, see Cache flash memory and Battery backup operations.
Cache capacity
The recommended amount of cache to install is determined by the RAID level, the number of drives installed in the system, and whether Hitachi Dynamic Provisioning (HDP), Hitachi Dynamic Tiering (HDT), Dynamic Cache Residency (DCR), and Universal Volume Manager (UVM) are applied. The recommended data cache capacity per Cache Logical Partition (CLPR) = (CLPR capacity) - (DCR Extent setting capacity per CLPR). When CLPR is not applied to DP/DT/DCR, install the recommended data cache capacity shown in the following table.
To configure a system for maximum performance, contact your authorized Hitachi Vantara representative. See Getting Help in the preface of this manual.
Total logical capacity of external volumes + internal volumes per CLPR | Recommended data cache capacity per CLPR |
Less than 2,900 GB |
15 GB or more |
Less than 2,900 GB |
15 GB or more |
2,900 GB or more |
16 GB or more |
11,500 GB or more |
22 GB or more |
14,400 GB or more |
24 GB or more |
100,000 GB or more |
30 GB or more |
128,000 GB or more |
32 GB or more |
182,000 GB or more |
40 GB or more |
218,000 GB or more |
48 GB or more |
254,000 GB or more |
56 GB or more |
290,000 GB or more |
64 GB or more |
326,000 GB or more |
72 GB or more |
Shared memory
Shared memory holds storage system configuration information and stored in the cache. The capacity of the shared memory + the capacity of the cache memory = the total capacity of the cache memory needed by the storage system.
The capacity overheads associated with the capacity saving function (data deduction) include capacity consumed by metadata and capacity consumed by garbage (invalid) data. For more information, see the Provisioning Guide for Open Systems. The recommendation is to use 0.2% of active data size as cache size (200 GB of cache for every 100 TB of pool capacity to be reduced).
The following table shows the shared memory capacity needed depending on the kind of software applications installed in the system.
Number of control unit | Determining factor of SM capacity |
iSCSI |
DC |
SM capacity | |||||||||
Software 1 |
64KLDEV extension |
HDP/ HDT/ AF extension | |||||||||||
SI/ VM/ NDM |
HDP/ TI/ FC |
TPF |
TC/ UR/ GAD |
HDT/ AF | 2 | 3 | 1 | ||||||
1-64 (16k LDEV) | Apply | Apply | — | — | — | — | — | — | — | — | — | 16 GB | |
1-64 (16k LDEV) | Apply | Apply | — | Apply | — | — | — | — | — | — | — | 24 GB | |
1-64 (16k LDEV) | Apply | Apply | — | — | Apply | — | — | — | — | — | — | 24 GB | |
1-64 (16k LDEV) | Apply | Apply | — | — | — | — | Apply | — | — | — | — | 24 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | — | Apply | — | — | — | Apply | Apply | 24 GB | |
1-64 (16k LDEV) | Apply | Apply | — | Apply | Apply | — | — | — | — | — | — | 32 GB | |
1-64 (16k LDEV) | Apply | Apply | — | Apply | — | — | Apply | — | — | — | — | 32 GB | |
1-64 (16k LDEV) | Apply | Apply | — | — | Apply | — | Apply | — | — | — | — | 32 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | — | Apply | — | — | — | Apply | Apply | 32 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | Apply | Apply | — | — | — | Apply | Apply | 32 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | Apply | Apply | — | — | Apply | Apply | 32 GB | ||
1-64 (16k LDEV) | Apply | Apply | — | Apply | Apply | — | Apply | — | — | — | — | 40 GB | |
1-64 (16k LDEV) | Apply | Apply | — | — | — | — | Apply | Apply | — | — | — | 40 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | Apply | Apply | — | — | — | Apply | Apply | 40 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | Apply | Apply | Apply | — | — | Apply | Apply | 40 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | — | Apply | Apply | — | — | Apply | Apply | 40 GB | |
1-64 (16k LDEV) | Apply | Apply | — | Apply | — | — | Apply | Apply | — | — | — | 48 GB | |
1-64 (16k LDEV) | Apply | Apply | — | — | Apply | — | Apply | Apply | — | — | — | 48 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | Apply | Apply | Apply | — | — | Apply | Apply | 48 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | — | Apply | Apply | Apply | — | Apply | Apply | 48 GB | |
1-64 (16k LDEV) | Apply | Apply | — | Apply | Apply | — | Apply | Apply | — | — | — | 56 GB | |
1-64 (16k LDEV) | Apply | Apply | — | — | — | — | Apply | Apply | Apply | — | — | 56 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | — | Apply | Apply | Apply | — | Apply | Apply | 56 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | Apply | Apply | Apply | Apply | — | Apply | Apply | 56 GB | |
1-64 (16k LDEV) | Apply | Apply | — | Apply | — | — | Apply | Apply | Apply | — | — | 64 GB | |
1-64 (16k LDEV) | Apply | Apply | — | — | Apply | — | Apply | Apply | Apply | — | — | 64 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | — | Apply | Apply | Apply | Apply | Apply | Apply | 64 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | Apply | Apply | Apply | Apply | — | Apply | Apply | 64 GB | |
1-64 (16k LDEV) | Apply | Apply | — | Apply | Apply | — | Apply | Apply | Apply | — | — | 72 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | — | Apply | Apply | Apply | Apply | Apply | Apply | 72 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | — | Apply | Apply | Apply | Apply | Apply | Apply | Apply | 72 GB | |
1-255 (64k LDEV) | Apply | Apply | Apply | Apply | Apply | Apply | Apply | Apply | Apply | Apply | Apply | 80 GB | |
|
Cache flash memory
The cache flash memory (CFM) is contained in the cache backup modules in the controller, along with the cache backup batteries. Similar to the cache memory, a CFM feature consists of a pair for redundancy purposes. The CFM backup the cache in case of power or component failure. The following figure shows two CPAs (2-3 and 1-4).
Cache flash memory operation
Each CFM blade connects directly to its corresponding CPA blade and backs up the data in that CPA blade if power fails. When data that is not stored on disk is written to the cache, it is written to one blade of the CPA and mirrored to the other. If one CFM fails, or if one phase of the power fails, the other CFM backs up the mirrored data from its corresponding CPA blade and data is not lost. In the unlikely event where a CFM has failed and a full power failure occurs, the other CFM backs up the mirrored data from the CPA without any loss of data.
Cache flash memory capacity
The recommended size of the installed cache flash memory depends on the size of the cache memory, and is automatically selected when defining the configuration for the system.
The following table shows CFM capacities per controller with both small and large cache memory backup assemblies.
Number of controllers |
Number of CFM features (pairs of boxes) |
Memory module size |
CFM size1 |
1 |
1 CFM 2 boxes / SSDs |
128 GB |
256 GB |
256 GB |
512 GB | ||
2 CFMs 4 boxes / SSDs |
128 GB |
512 GB | |
256 GB |
1 TB | ||
2 |
1 CFM 2 boxes / SSDs |
128 GB |
512 GB |
256 GB |
1 TB | ||
2 CFMs 4 boxes / SSDs |
128 GB |
1 TB | |
256 GB |
2 TB | ||
Notes:
|
Service Processor
The service processor (SVP) is a hardware component that performs modifications of settings, reports statistical information about device availability, and provides maintenance accessibility to the storage system.
The controller chassis is equipped with a primary and an optional secondary service processor. The primary SVP is active while the secondary SVP is duplicated and remains on active-standby in case of a failure. Employing a duplicate SVP configuration provides redundancy and prevents the loss of use of its monitoring function. If a failure occurs, the standby SVP is automatically switched into operation. The switching operation time is approximately three minutes.
The table lists the technical specifications of the service processor.
Component |
Specifications |
Operating System |
Windows 7 / Windows 10 IoT Enterprise* |
CPU |
Intel Celeron P4505 1.86 GHz |
Internal Memory |
4 GB |
Disk drive |
300 GB (3.5-inch HDD) |
LAN |
On-Board 10Base-T/ 100Base-TX /1000Base-T x 2 Port |
HUB |
On-Board 10Base-T/ 100Base-TX /1000Base-T x 19 Port |
Modem |
None |
Serial port |
RS-232-C |
USB |
Version 2 x 4 ports |
PC Card Slot |
None |
*The SVP must be on firmware version 80-06-42 or later to support the Windows 10 IoT Enterprise operating system. |