• Topic ID: id_15460905
  • Version: 2.0
  • Date: Nov 8, 2018 1:38:01 AM

VDARC Replacement or Reload

Prerequisites

Overview

This module describes the procedures used to replace the VDARC Node (Westville or Jarrell version) in the GOC5 VCT Operator Console. After the VDARC Node has been replaced and loaded with the OS and Apps software, several tests are required:

  • Communication and verification within the VDARC Node

  • Communication to the VIG Nodes

  • Communication to the Disk Array

  • Recon Data Path Diagnostic

  • Resave of the System State Reconfig Info

  • Sanity Scans

note:

Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent.

note:

Effective software version 12HW14.6, DO NOT perform Sections 4.8.

Instead a complete Load From Cold is required (including Host OS and APPS loading)!

Go to software chapter in this service manual and follow the LFC instructions for 12HW14.6.

Upon completion of LFC, Sections 4.9 – 4.14 may be performed to check for proper functionality after DARC replacement.

1 Additional Software Supporting the VDARC Node

2 Power-OFF (Shut Down) the Operator Console

Procedure

  1. Select one of the following methods to Power OFF the Operator Console:
    1. If Applications are up, select the Shutdown icon, then select Shutdown.
    2. Notice Image
      • notice
      • (For 06MW03.X Software) Wait 1-2 minutes after the “System halted” message appears, to allow the DARC Node to properly power-down. Future SW releases (e.g., 06MW29.7) have corrected this issue.
    3. If Applications are down, open a Unix Shell at the Toolchest. Type:

      {ctuser@ hostname }   halt

    The Operator Console monitor will display a System halted message when it is acceptable to power OFF the Operator Console.

  2. Power OFF the Operator Console at the front panel switch. See Figure 1.
    note:

    Do not turn OFF the VDARC Rear Power Switch.

    Figure 1. Operator Console Power Switch

3 Remove the Old VDARC

Procedure

  1. Remove the front and rear Operator Console covers.
  2. Verify each cable label is present and clearly marked. If necessary, add a label for clarity.
  3. Disconnect all rear VDARC Node cables.

    Refer to Table 6 for cables to be removed. (Please refer to Figure 2 or Figure 3 below for VDARC Node Types and connection information.)

    Figure 2. “Jarrell” VDARC Node #5159833

    Sheet 1 of 2

    Sheet 2 of 2

    Figure 3. “Westville” VDARC Node #5114523

    Sheet 1 of 2

    Sheet 2 of 2

  4. Remove and retain the four (4) front panel mounting screws.
  5. Slide the VDARC Node out of the Operator Console from the front.

4 Install the New VDARC

Refer to 5115335SCH - GOC5 Interconnect, as required.

Procedure

  1. Slide the replacement VDARC Node into the Operator Console from the front.
  2. Replace the (4) front mounting screws on the front panel of the VDARC Node.
  3. Reconnect the VDARC Node cables. See Figure 2 or Figure 3, depending on VDARC Node version.
  4. At the rear panel of VDARC Node, verify the power switch is in the “ON” position.

5 Power ON the Operator Console

Procedure

  1. Power ON the Operator Console at the front power switch.
  2. Verify the VDARC Node front panel power LED is ON. If not, press the front panel power switch (hold 3-10 seconds) to apply power to the VDARC Node.
  3. Verify that the VIG Node front panel power LED is illuminated. If not, press the front panel power switch (hold 3-10 seconds) to apply power to the VIG Node. See Figure 4 or Figure 5, depending on VIG Node version.

    Figure 4. “Jarrell” VIG Node #5159834

    Sheet 1 of 3

    Sheet 2 of 3

    Sheet 3 of 3

    Figure 5. ”Westville” VIG Node #5114533-2

    Sheet 1 of 3

    Sheet 2 of 3

    Sheet 3 of 3

  4. At the Disk Array (JBOD), verify the two green LEDs at the rear panel are illuminated. If they are not, then press the two power switches at the rear panel, to apply power to the Disk Array. See Figure 6.

    Figure 6. Rear Panel of the VCT Disk Array

  5. Stop Applications software from auto-starting at the 5-second Pink Box.

6 Verify VDARC Node Is Ready for Software Load

The VDARC Node must be UP, to load the OS (Operating System) and DARC Application software. Perform the telnet command on the VDARC Node to verify communication, as follows:

Procedure

  1. At the Toolchest, with Application software down, open a Unix Shell.
  2. Perform the telnet command to the VDARC Node from the Host Computer:

    {ctuser@hostname}   su -

    Password:   #bigguy

    [root@hostname]   service  cliservice  start

    dpcproxy server will be reported as already running or restarted

    [root@hostname]:  telnet  localhost  623

    Trying 127.0.0.1...

    Connected to localhost

    *** Indicates cliservice proxy server is running ***

    Escape character is '^]'.

    Server:   darc

    Username:   Enter

    Password:   Enter

    Login successful *** Indicates the SOL connection is established ***

    dpccli>   exit

    Connection closed by foreign host.

    [root@hostname]

    note:

    If the telnet connection fails, refer to Telnet T/S Note.

7 Confirm Host PC Software

Confirm the Host Computer Software type. Examples are shown, the software version output should match the Software media to be loaded.

Procedure

  1. Open a Unix Shell.
  2. Type the following to verify software version information. Examples are provided. The examples are not actual output.
    note:

    Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent

    1. {ctuser@ hostname }   swhwinfo   << for Apps Version

      (year) MW (fiscal week).(fiscal day).hardware revision info here

      .Example: 06MW29.7.V40_H_V64_G_GTL

      note:

      If the revisions do NOT match the software media to be loaded, locate the proper software set. If the software set cannot be found or ordered, a full LFC will be required for the Operator Console.

      If the revisions match the software media to be loaded, continue with the load procedure.

    2. {ctuser@ hostname }   swhwinfo  -o   << for OS version

      Release: GEHC/CTT Linux 4.3.16

      Built: Tue Jul 12 12:29:10 CTD 2005

      note:

      Confirm that the results match the Operating System version and Software Build Date shown on the OS and Applications DVDs.

      Effective software version 12HW14.6, DO NOT perform Sections 4.8.

      Instead a complete Load From Cold is required (including Host OS and APPS loading)!

      Go to software chapter in this service manual and follow the LFC instructions for 12HW14.6.

      Upon completion of LFC, Sections 4.9 – 4.14 may be performed to check for proper functionality after DARC replacement.

8 Load Software from Host onto VDARC Node

8.1 Standard VDARC Load Procedure

The VDARC Node must be loaded with the OS and Application software from the Host Computer to the VDARC Node. Perform the VDARC Node OS (Operating System) load as described in the LFC (Load From Cold) procedure for the software set being used. Perform the VDARC Node Application software load as described in the LFC procedure for the software set being used. Refer to the appropriate LFC procedure and follow the sub-sections after reading the DARC Subnet information outlined below:

note:

Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent

note:

The OS (Operating System) and Application software version loaded onto the VDARC Node must match the OS and Application software versions loaded on the Host Computer.

Procedure

  1. Perform the VDARC Node OS (Operating System) load as described in the LFC (Load From Cold) procedure for the software set being used.
    note:

    If the OS software fails to load successfully on the VDARC Node from the Host Computer, refer to OS Load Work-Around with Disk Array (JBOD) OFF for work-around procedure and VDARC Stand-Alone OS Load procedure.

    • Refer to the “LFC - DVD OS Installation on VDARC” section of the appropriate 7.X LFC Procedure.

      note:

      Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent

    • Perform an OS (LS VCT Console Operating System DVD) load on the VDARC Node from the Host Computer per the normal instructions.

  2. Perform the VDARC Node Application load as described in the LFC (Load From Cold) procedure for the software set being used.
    note:

    If the DARC Application software fails to load successfully on the VDARC Node from the Host Computer, refer to DARC Apps Load Work-Around with Disk Array (JBOD) OFF for work-around procedure.

    • Refer to the “LFC - CD-ROM VDARC Application Installation on Host Computer” section of the appropriate 7.x LFC Procedures.

    • Perform a VDARC Applications software load on the VDARC Node from the Host Computer per the normal instructions. Refer to the DARC Subnet Note below.

    note:

    DARC Subnet Address Information: The following procedure should be performed if your customer Ethernet hospital backbone begins with 172.16.0.xx, OR if the scanner will connect to a device with an IP address of 172.16.0.xx.

    • If the hospital backbone starts with 172.16.0.xx, and the VDARC subnet was previously set to 169.254.0 during the LFC procedure on this Operating Console, verify that the Change DARC subnet box is checked during the Application software load and the DARC subnet address is set at 169.254.0.

    • If the VDARC subnet was previously set to 169.254.0 and needs to be set back to the 172.16.0 subnet, uncheck the Change DARC subnet box during the Application software load

  3. If the VDARC Node software has completed successfully, skip to Verify Replacement VDARC Node to verify the VDARC Node.

8.2 OS Load Work-Around with Disk Array (JBOD) OFF

If any problems arise during the VDARC Node OS software load, verify the correct media is installed or verify the Boot BIOS Order is correct. If VDARC Node OS load problems continue, perform the OS Load Work-around with Disk Array (JBOD) OFF. If problems still continue, perform the VDARC Stand-Alone OS Load procedure. The stand-alone procedure eliminates any external sources causing an issue but does not rule out media issues or internal VDARC Node issues.

note:

Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent

note:

If VDARC Node OS load failed, perform the OS Load Work-around with Disk Array (JBOD) OFF. Do Not perform this section if the VDARC Node has loaded the OS from the Host Computer successfully.

Procedure

  1. Remove the rear cover from the Operator Console.
  2. At the rear of the Disk Array (JBOD), locate the two green Power LEDs at rear panel. Press the two (2) power switches at rear panel to remove power from the Disk Array. See Figure 7.

    Figure 7. Rear Panel of VCT Disk Array

  3. Repeat the steps in Standard VDARC Load Procedure.
    note:

    Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent

  4. If this work-around fails, perform the VDARC Stand-Alone OS Load per the following Section.
  5. If this work-around is successful, at the Disk Array (JBOD) verify two green LEDs at rear panel are illuminated. If not, press two power switches at rear panel, to apply power to the Disk Array. See Figure 7.

8.3 DARC Apps Load Work-Around with Disk Array (JBOD) OFF

If any problems arise during the VDARC Node Apps software load, verify the correct media is installed or verify the Boot BIOS Order is correct. Remove the rear cover from the Operator Console.

note:

Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent

note:

If VDARC Node Apps Load failed, perform the DARC Node Load Work-around with Disk Array (JBOD) OFF. Do Not perform this section if the VDARC Node has loaded the Apps software successfully from the Host Computer.

Procedure

  1. Verify the OS has been loaded on the VDARC Node, open a Unix Shell and type the following:
    1. {ctuser@ hostname }  su -
    2. Password:   #bigguy
    3. [root@ hostname ]   rsh darc
    4. [root@ localhost root ]  exit
    note:

    If the VDARC login prompt [root@localhost root] is not present, then try reloading the VDARC Node OS again per the procedure. Always be sure the VDARC OS matches the Host OS already loaded. If issues persist perform the VDARC Stand-Alone OS Load procedure.

  2. At the rear of the Disk Array (JBOD), locate the two green Power LEDs at rear panel. Press the two (2) power switches at rear panel to remove power from the Disk Array. See Figure 8

    Figure 8. Rear Panel of VCT Disk Array

  3. Repeat the steps in Standard VDARC Load Procedure, step 2. Perform the VDARC Node Application load as described in the LFC (Load From Cold) procedure for the software set being used.
    note:

    Jarrell VDARC and Jarrell IG Nodes require software release 06MW29.7 or more recent

    note:

    During the VDARC Apps software load, process errors may occur directly related to the Disk Array power off state. Ignore these errrors.

  4. If this DARC Apps load work-around is successful, at the rear panel of the Disk Array (JBOD) press two power switches to apply power to the Disk Array. See Figure 8.
  5. After the VDARC Node has completed rebooting, verify the DARC Apps software has been loaded on the VDARC Node. Open a Unix Shell and type the following:
    1. {ctuser@ hostname }   su -
    2. Password:   #bigguy
    3. [root@hostname]   rsh darc
    4. [root@darc]   exit
    note:

    If the VDARC login prompt [root@darc] is not present, then the work-around has not been successful. Verify the telnet process from the Host to the VDARC Node is working and retry the load procedure.

  6. When the DARC Application Software is loaded with the Disk Array off, the Disk Array drivers are not loaded. Reboot the console at this time.
  7. Once the system has rebooted, wait for the VDARC to come up.
  8. Perform the following commands to repartition/create the Disk Array. Verify 8 disks are present. Visually verify the entire output is successful from the command script performed. The final output line ‘gre-raid success’ does not mean the Disk Array is good.
    1. {ctuser@ hostname }   rsh darc
    2. {ctuser@darc}   sudo gre-raid -z  -c
    3. answer yes
    4. {ctuser@darc}   sudo gre-raid -a
    5. {ctuser@darc}   sudo gre-raid -q
    note:

    If the VDARC login prompt [root@darc] is not present, then the work-around has not been successful. Verify the telnet process from the Host to the VDARC Node is working and retry the load procedure.

9 Verify Replacement VDARC Node

With Application Software Down, the VDARC Node must be checked to verify communication between the Host Computer and the VDARC Node. Perform ping and rsh commands on the VDARC Node, as follows:

note:

If the ping/rsh commands fail, refer to VDARC Port Communication T/S Note.

Procedure

  1. At the Toolchest, with Application software down, open a Unix Shell.
  2. Ping the VDARC Node:

    {ctuser@ hostname }   ping darc

    Verify ping is successful.

    Output similar to the following will appear:


    PING darc (172.16.0.2) from 172.16.0.1 : 56(84) bytes of data.
    64 bytes from darc (172.16.0.2): icmp_seq=1 ttl=64 time=0.157 ms
                    

    Press Ctrl+C to stop ping process.

    Output similar to the following will appear:


    --- darc ping statistics ---
    4 packets transmitted, 4 received, 0% loss,time 3000ms
    rtt min/avg/max/mdev = 0.134/0.168/0.209/0.031 ms
                    

  3. Remote Shell to the VDARC Node:

    {ctuser@ hostname }   rsh darc


           Last login: Thu Oct 9 11:17:35 on ttyS1
           You have new mail.
    {ctuser@darc}
                    

    Verify ctuser@darc prompt is present.

  4. (For 06MW03.x SW and greater) Security Shell to the VDARC Node:

    {ctuser@ hostname }   ssh darc


           Last login: Thu Oct 9 11:17:35 on ttyS1
           You have new mail.
    {ctuser@darc}
                    

    Verify ctuser@darc prompt is present. If not, a reconfig (as root on the host) and Accept will be required to sync up the known_host files.

  5. Exit from the DARC and become root.
  6. Type:  rsh  darc
  7. As root, type:  vdip_menu  -A
  8. Verify all tests pass.
    note:

    The Unix Shell will be used again, so do not exit or close at this time.

10 Apply Power to VIG Nodes

Remove the power for each VIG Node, then re-power each VIG Node with the VDARC Node up and able to Remote Shell (rsh darc).

Procedure

  1. At each VIG Node, verify the power is OFF. If power is On for any VIG Node, then press the front panel power switch (hold 3-10 seconds) to remove power to the VIG Node(s).
  2. Wait 20 seconds.
  3. At each VIG Node, press the front panel power switch (hold 3-10 seconds) to apply power to the VIG Node(s).
  4. Wait two (2) minutes for the VIG Nodes to boot up.

11 Verify Communication from VDARC Node to VIG Nodes

The VDARC Node must be up to check the VIG Node(s). Perform ping and rsh commands on the VDARC Node, as follows:

note:

If the VIG Node ping/rsh commands fail, refer to VIG Port Communcation T/S Note.

Procedure

  1. At the Toolchest, with Application software down, if necessary, open a Unix Shell.
  2. Ping VIG Node #1:

    {ctuser@darc}   ping ig1

    Verify ping is successful.

    Press Ctrl+C to stop ping process.

  3. Remote Shell to the VIG Node #1:

    {ctuser@darc}   rsh ig1

    Verify ctuser@ig1 prompt is present.

  4. Exit from the VIG Node #1 Remote Shell:

    {ctuser@ig1}   exit

  5. Ping VIG Node #2:

    {ctuser@darc}   ping ig2

    Verify ping is successful.

    Press Ctrl+C to stop ping process.

  6. Remote Shell to the VIG Node #2:

    {ctuser@darc}   rsh ig2

    Verify ctuser@ig2 prompt is present.

  7. Exit from the VIG Node #2 Remote Shell:

    {ctuser@ig2}   exit

  8. Ping VIG Node #3:

    {ctuser@darc}   ping ig3

    Verify ping is successful.

    Press Ctrl+C to stop ping process.

  9. Remote Shell to the VIG Node #3:

    {ctuser@darc}   rsh ig3

    Verify ctuser@ig3 prompt is present.

  10. Exit from the VIG Node #3 Remote Shell:

    {ctuser@ig3}   exit

  11. Exit from the VDARC Node Remote Shell:

    {ctuser@darc}   exit

    {ctuser@ hostname }

12 Verify Disk Array

In the following test, the Disk Array is configured with 8 disks (4 on each side). Visually verify the entire output is successful from the command script performed. The final output line ‘gre-raid success’ does not mean the Disk Array is good.

note:

Application Software must be down to perform the assemble command.

Procedure

  1. {ctuser@ hostname }   rsh darc
  2. {ctuser@darc}   sudo gre-raid -a
  3. Verify assemble diagnostic passes and all output looks good.
  4. {ctuser@darc}   sudo gre-raid -q
  5. Verify query diagnostic passes and all output looks good.
  6. {ctuser@darc}   exit
  7. {ctuser@ hostname }
    note:

    The Unix Shell will be used again, so do not exit or close at this time.

13 Start Application Software

{ctuser@ hostname }   st

14 Verify Recon Data Path Diagnostic

Procedure

  1. Select the Service icon.
  2. Select DIAGNOSTICS
  3. Select Recon Data Path
  4. Select 1 and ALL
  5. Select RUN
  6. Verify Recon Data Path tests pass.
  7. Select Dismiss

15 Resave Reconfig Info File on System State

After the VDARC Node software load (OS and Apps), the Reconfig Info file must be resaved on the System State DVD-RAM. NOTE: Effective software release 12HW14.6, this step may be skipped. Proceed to Finalization section. System State Save will be accomplished during the LFC process.

Procedure

  1. Insert the System State DVD-RAM into the SCSI Tower DVD RAM drive.
  2. Wait until the DVD drive is ready (i.e., until the front panel green LED is no longer lit).
  3. Select the Service icon to access the CSD (Common Service Desktop).
  4. Select Utilities
  5. Select System State

    Figure 9. System State GUI

  6. Select Reconfig Info
  7. Select Save.
  8. If applicable, select [Yes] when the following message appears: System State Media Status: Please insert a DVD into the drive and press Save again.

    Figure 10. Save System State DVD Pop-Up

    note:

    Verify that the "Save" of System State was successful. If not, correct any errors and re-save the Reconfig Info file on the System State DVD-RAM. A message at the end of the Save should state: Save/Restore System State: Completed Successfully.

    Figure 11. Save System Message Log (Example Only)

  9. When “Save/Restore System State: Completed Successfully.” message appears, select Cancel
  10. When completed, select Dismiss

    IMPORTANT: Do NOT select Dismiss more than once. Doing so will result in the GUI becoming hung, and the System State will be unusable. Should this occur, Save the System State again.

  11. Close the Service Desktop window at the upper left corner of the screen.

16 Finalization

Procedure

  1. Perform sanity scans (e.g., scout, axial and helical).
    note:

    If the System will not scan or reconstruct verify the memory on the VDARC Node is correct (lhinv). Verify the no loop-back VDIP diagnostic passes (vdip_menu –A). Verify the VRAC diagnostics pass (performed as root: vrac_menu –a).

  2. Verify the image(s) reconstruct and are properly displayed.
  3. Install the front and rear Operator Console covers.