cancel
Showing results for 
Search instead for 
Did you mean: 
cancel
781
Views
5
Helpful
7
Replies
Highlighted
Beginner

Specifying file path for config Archive causes stack reboot.

 When specifying a file path for our config archive, the standby and member switches reboot. While the active switch stays up, the other switches in the stack reboot as soon as the command is entered. If the command is not removed before the switches complete their reboot sequence, they boot into ROMmon.

Is this expected behavior, config issue or bug? Any help is greatly appreciated. Thanks!

 

Hardware and versions of tested stacks.

Bootloader: 16.12.2r

IOS:16.12.2 and 16.9.4

9300-48A-UXM

9300-48A-P

 

Here  are the commands entered:

(config)# archive

(config-archive)# path flash:/Netops/Rollback/

 

Prior to entering the commands:

Switch# Role Mac Address Priority Version State
-------------------------------------------------------------------------------------
*1 Active XXXX.XXXX.XXXX 15 V02 Ready
2 Standby XXXX.XXXX.XXXX 5 V02 Ready
3 Member XXXX.XXXX.XXXX 1 V02 Ready

 

Immediately After:

Labs-1FL-SS#sh switch
Switch/Stack Mac Address : XXXX.XXXX.XXXX - Local Mac Address
Mac persistency wait time: Indefinite
H/W Current
Switch# Role Mac Address Priority Version State
-------------------------------------------------------------------------------------
*1 Active XXXX.XXXX.XXXX 15 V02 Ready
2 Member 0000.0000.0000 0 V02 Removed
3 Member 0000.0000.0000 0 V02 Removed

 

1 ACCEPTED SOLUTION

Accepted Solutions
Highlighted

hmm thats a good workaround but that would suggest somethings definitely up with the software version , we have a lot of c9300 stacks running 16.6.6 and we have not seen that issue , archive and kron running on all of them , any bug that reboots a switch you should move off that version if possible

View solution in original post

7 REPLIES 7
Highlighted
VIP Mentor

Hi
Is there any log before it reboots ?

neither release doc currently identify a bug id that matches that so it may not be known , so you will need to go to TAC or else try eh Cisco analyzer see if that gives you a bug id , or else test another image move off Gibraltar and Fuji , try Amsterdam

https://www.cisco.com/c/en/us/td/docs/switches/lan/catalyst9300/software/release/16-12/release_notes/ol-16-12-9300.html

https://www.cisco.com/c/en/us/td/docs/switches/lan/catalyst9300/software/release/16-9/release_notes/ol-16-9-9300.html
Highlighted

Also is that a typo in how your running the commands as it should be path flash: just to be sure its not that
Highlighted

Thanks. Yes just typo on the post here. 

 

Highlighted

Yeah I could not find any bugs. Here is the log output.

 

001288: .Jan 9 21:03:04.818: Config Sync: Bulk-sync failure due to PRC mismatch. Please check the full list of PRC failures via:
show redundancy config-sync failures prc

001289: .Jan 9 21:03:04.818: Config Sync: Starting lines from PRC file:
archive
! <submode> "archive"
- path flash:/NetOps/Rollback
! </submode> "archive"

001290: .Jan 9 21:03:04.818: Config Sync: Bulk-sync failure, Reloading Standby

001291: .Jan 9 21:03:05.825: %RF-5-RF_TERMINAL_STATE: Terminal state reached for (SSO)
001292: .Jan 9 21:03:06.274: %RF-5-RF_RELOAD: Peer reload. Reason: Bulk Sync Failure
001293: .Jan 9 21:03:06.644: %HMANRP-5-CHASSIS_DOWN_EVENT: Chassis 3 gone DOWN!
001294: .Jan 9 21:03:06.657: %REDUNDANCY-3-STANDBY_LOST: Standby processor fault (PEER_NOT_PRESENT)
001295: .Jan 9 21:03:06.657: %REDUNDANCY-3-STANDBY_LOST: Standby processor fault (PEER_DOWN)
001296: .Jan 9 21:03:06.657: %REDUNDANCY-3-STANDBY_LOST: Standby processor fault (PEER_REDUNDANCY_STATE_CHANGE)
001297: .Jan 9 21:03:06.578: %STACKMGR-6-STACK_LINK_CHANGE: Switch 1 R0/0: stack_mgr: Stack port 2 on Switch 1 is down
001298: .Jan 9 21:03:06.620: %STACKMGR-6-STACK_LINK_CHANGE: Switch 2 R0/0: stack_mgr: Stack port 1 on Switch 2 is down
001299: .Jan 9 21:03:08.149: %RF-5-RF_RELOAD: Peer reload. Reason: EHSA standby down

Highlighted

My next step would be collect the show tech and run it yourself through the Cisco cli analyser available in tools section of website see if it can pinpoint an issue , if not open a TAC case
Also check each flash has not generated a crash file , the log output unfortunately only shows it doing a standard down log , not a trigger that may have caused it
Highlighted
Beginner

Hi,

 

I just had this issue.

 

What i did and it seems to fix the issue was to create the archive folder on each switches flash.

If i just created it just on flash: it would reboot the other members. Since the other stack member cant save it to its local directory as it doesn't exist.

 

Try creating the directories on the other switches.

example

mkdir flash-1:Netops/Rollback/

mkdir flash-2:Netops/Rollback/

mkdir flash-2:Netops/Rollback/

 

Highlighted

hmm thats a good workaround but that would suggest somethings definitely up with the software version , we have a lot of c9300 stacks running 16.6.6 and we have not seen that issue , archive and kron running on all of them , any bug that reboots a switch you should move off that version if possible

View solution in original post

Content for Community-Ad