IC SunsetThe developerWorks Connections platform will be sunset on December 31, 2019. On January 1, 2020, this blog will no longer be available. More details available on our FAQ.

Comments (18)
  • Add a Comment
  • Edit
  • More Actions v
  • Quarantine this Entry

Comments (18)

purdym commented Dec 12 2015 Comment Permalink

Good article. For me the solution was the firewall rules on the interface in the HMC. Ensure RMC is allowed through the HMC's firewall.
hmc management -> change network setting -> lan adapter -> choose adapter -> details
-> firewall settings -> select RMC -> 'allow incoming'

Ensure RMC is now under 'Allowed hosts'.

Musfar commented May 7 2015 Comment Permalink

I ran into the problem recently. And after trying several things, we narrowed down to the network restrictions between the HMC subnet and LPARs. Learnt that the subnet of HMC is a general computing, while the LPARs' subnet is restricted. So, we requested the firewall team to allow HMC IP over port 657 to access restricted subnet of the LPARs. Hope this will be of some help to others.

M. Veidt commented Feb 11 2015 Comment Permalink

Hi Chris, great post. Also a common reason for RMC not working: HMC hostname is not resolvable via DNS or local lookup from the client LPAR. This should explicitly checked a well.

POWERHAguy commented Feb 6 2015 Comment Permalink

I did prove rcotter comment about running recfgct does indeed crash a PowerHA node cluster. So as my original notes state, take powerha and caa down before hand.

POWERHAguy commented Jan 29 2015 Comment Permalink

I just encountered a similar problem for cloning a primary node in PowerHA and making it a standby node. Though the environment as far as PowerHA was concerned worked, but errpt every 60 seconds was recording the errors I listed below. I went through the procedures above, though I stopped PowerHA on the standby node before starting. After performing the steps I got cthags was no longer found. I rebooted and all worked but I wasn't real happy with that.
So I recreated my environment again. The only step difference this time was when stopping PowerHA on the stby node I also stopped CAA services, went through steps, then restarted and told it to start CAA services again. Now this exact clmgr syntax only works with PowerHA 7.1.3 SP1 or above. Earlier versions of CAA/HA have different options/commands to stop/start CAA individually. This seems to have worked for me, hopefully it works for others. clmgr stop node dtcu0_stby WHEN=now MANAGE=offline STOP_CAA=yes stopsrc -g rsct_rm; stopsrc -g rsct /usr/bin/odmdelete -o CuAt -q 'attribute=node_uuid' /usr/sbin/rsct/bin/mknodeid -f (when I ran this step I go no output, I think it just pulls in the existing one from repos disk for node but not sure) lsattr -El cluster0 /usr/sbin/rsct/bin/lsnodeid /usr/sbin/rsct/install/bin/recfgct clmgr start node web WHEN=now MANAGE=auto START_CAA=yes ________________________________________________________________________________________- LABEL: CONFIGRM_ONLINEFAIL IDENTIFIER: E509DBCA LABEL: CONFIGRM_STARTED_ST IDENTIFIER: DE84C4DB LABEL: SRC_RSTRT IDENTIFIER: CB4A951F LABEL: CONFIGRM_EXIT_ONLIN IDENTIFIER: 68FD23E8

Rick Cotter commented Mar 22 2014 Comment Permalink

Warning: The recfgct command referenced above is *not* supported for use by customers without direct IBM support instructions. It erases all RSCT configuration info and makes it look like the node was just installed. This may be fine for DLPAR recycling, but if you have any other products dependent on RSCT on the partition in question, you will be *broken*. In particular, PowerHA 7 will crash, and Tivoli SAMP will have all its cluster info destroyed, partitioning it from the rest of the domain until it can be manually re-added (and it may also crash, depending on the presence of resources). If you find that DLPAR is not working, and all other network checks and even the RMC recycling (-z/-A/-p) does not work, it is strongly recommended that you use the ctsnap command to gather data and contact IBM support. (Capturing iptrace for a few minutes would not be a bad idea either. A complementary tcpdump on the HMC would also be good, but this may not be possible for most customers given HMC's access restrictions.) Then, if you wish to proceed with recfgct and find that it does resolve whatever the problem was, it would be equally wise to gather another ctsnap after the partition is once again connected to the HMC, to compare to the previous one.

cggibbo commented Feb 12 2014 Comment Permalink

Sounds looks like you might have a "ghost" adapters info on the HMC and in the VIOS. So now things are out of sync. You could try the following (at your own risk), to resolve the problem: From oem_setup_env on the VIOS: # /usr/sbin/drmgr -a -c slot -s U911X.MXX.1234E8C-V1-C164 -d 5 ; where the location code matches your adapter/slot config. Reconfigure the slot in the VIOS from padmin: $ cfgdev If the above works as expected then you should be able to remove the VFC adapter now, as padmin: $ rmdev -dev vfchostXYZ Then the HMC DLPAR remove on the slot should complete and leave the HMC and VIOS partition in a consistent state.

Nolte commented Feb 11 2014 Comment Permalink

Hi Chris, thank for the article very useful. But i have a problem: Adding in a dlpar mode a virtual fiber channel adapter to my vio, i have received an error in communication, but i have clicked "OK" in the window. After reset the connection following this article, the problem is that is not possible to remove the adaper from the vio running profile because : "0931-009 You specified a drc_name for a resource which is not assigned to this partition." effectively i don't have the vfchost on vio..(not even in "unknow state"). is like an error from HMC and a my error clicking "OK" and not "cancel" when the communication is bad. obviously after the reset is possible to make dlpar actions but is not possible delete the adapter even whit --force option in chhwres command. Thanks.

VEUT_xu_ma commented Mar 12 2013 Comment Permalink

Thanks for your article. I resolve my problem.

hillanes commented Nov 21 2012 Comment Permalink

Thks Chris, What about IVM environment, without HMC. I cant start RMC service. How can I troubleshoot my environment?