Natwest/RBS glitch
Discussion
ExFiF said:
joe_90 said:
nail.. head.
The best thing is they just say yes.. or go quiet and do/say nothing.
Actually I'd go further and say that they say everything is OK< or go quiet even when pressed repeatedly. When the situation is about to go into meltdown and they are threatened with their intransigence / failure to reply being escalated up the management tree, then their reply pings back within seconds. This says to me that they did know the answer but were just refusing to answer.The best thing is they just say yes.. or go quiet and do/say nothing.
Of course if you do escalate it and their boss is of same mindset...
Sexual Chocolate said:
Nothing to do with change control. Its pretty strict here.
The upgrade was succsfull, it just had performance issues. By monday they where about a day behind. On Tuesday when they backed it out someone formatted the messaging queue. Not sure what they thought was going to happen or if this was done in error. Could have been as simple as the UNIX/LINUX equvilant of sudo rm -rf * but not checking where you where in the system.
The change was implemented by UK based staff the recovery on the other hand wasn't.
so we broke it, and the Indian's fixed it??The upgrade was succsfull, it just had performance issues. By monday they where about a day behind. On Tuesday when they backed it out someone formatted the messaging queue. Not sure what they thought was going to happen or if this was done in error. Could have been as simple as the UNIX/LINUX equvilant of sudo rm -rf * but not checking where you where in the system.
The change was implemented by UK based staff the recovery on the other hand wasn't.
Edited by Sexual Chocolate on Thursday 28th June 10:34
onyx39 said:
Sexual Chocolate said:
Nothing to do with change control. Its pretty strict here.
The upgrade was succsfull, it just had performance issues. By monday they where about a day behind. On Tuesday when they backed it out someone formatted the messaging queue. Not sure what they thought was going to happen or if this was done in error. Could have been as simple as the UNIX/LINUX equvilant of sudo rm -rf * but not checking where you where in the system.
The change was implemented by UK based staff the recovery on the other hand wasn't.
so we broke it, and the Indian's fixed it??The upgrade was succsfull, it just had performance issues. By monday they where about a day behind. On Tuesday when they backed it out someone formatted the messaging queue. Not sure what they thought was going to happen or if this was done in error. Could have been as simple as the UNIX/LINUX equvilant of sudo rm -rf * but not checking where you where in the system.
The change was implemented by UK based staff the recovery on the other hand wasn't.
Edited by Sexual Chocolate on Thursday 28th June 10:34
For some reason that I am not sure of they had to go back to the Friday back-up which then led to the chaos as all accounts would be brought back to the Friday position and all transactions since this point would have to be manually restored.
Not sure of they did restore from Friday there is no mention of it. But then again I don't know mainframe stuff so maybe they did.
Reading through it all the change requestor mentions its a simple task and hence thats why they got that window. Probably implelemented by prod support which is, as far as I know is offhsore but they are still RBS employees I think.
Reading through it all the change requestor mentions its a simple task and hence thats why they got that window. Probably implelemented by prod support which is, as far as I know is offhsore but they are still RBS employees I think.
Edited by Sexual Chocolate on Thursday 28th June 11:19
Chim said:
onyx39 said:
Sexual Chocolate said:
Nothing to do with change control. Its pretty strict here.
The upgrade was succsfull, it just had performance issues. By monday they where about a day behind. On Tuesday when they backed it out someone formatted the messaging queue. Not sure what they thought was going to happen or if this was done in error. Could have been as simple as the UNIX/LINUX equvilant of sudo rm -rf * but not checking where you where in the system.
The change was implemented by UK based staff the recovery on the other hand wasn't.
so we broke it, and the Indian's fixed it??The upgrade was succsfull, it just had performance issues. By monday they where about a day behind. On Tuesday when they backed it out someone formatted the messaging queue. Not sure what they thought was going to happen or if this was done in error. Could have been as simple as the UNIX/LINUX equvilant of sudo rm -rf * but not checking where you where in the system.
The change was implemented by UK based staff the recovery on the other hand wasn't.
Edited by Sexual Chocolate on Thursday 28th June 10:34
For some reason that I am not sure of they had to go back to the Friday back-up which then led to the chaos as all accounts would be brought back to the Friday position and all transactions since this point would have to be manually restored.
honest_delboy said:
Chim said:
From what I can gather, yes. Only mystery to me is why they had to restore from Friday.
With the scheduler offline to do the upgrade ..... would the scheduler schedule its own backups? joe_90 said:
ExFiF said:
joe_90 said:
nail.. head.
The best thing is they just say yes.. or go quiet and do/say nothing.
Actually I'd go further and say that they say everything is OK< or go quiet even when pressed repeatedly. When the situation is about to go into meltdown and they are threatened with their intransigence / failure to reply being escalated up the management tree, then their reply pings back within seconds. This says to me that they did know the answer but were just refusing to answer.The best thing is they just say yes.. or go quiet and do/say nothing.
Of course if you do escalate it and their boss is of same mindset...
But agree it is a cultural thing, also having said what written earlier, some Indian colleagues who have come to Europe have been some of the most committed, hard working, never say die colleagues I have ever had the privilege to work with.
Sexual Chocolate said:
onyx39 said:
so we broke it, and the Indian's fixed it??
Change raised by uk staff, issue discovered by UK based staff (Ops I think) the indian support team didnt have a clue how to fix it. In the end UK staff resolved it.ExFiF said:
joe_90 said:
ExFiF said:
joe_90 said:
nail.. head.
The best thing is they just say yes.. or go quiet and do/say nothing.
Actually I'd go further and say that they say everything is OK< or go quiet even when pressed repeatedly. When the situation is about to go into meltdown and they are threatened with their intransigence / failure to reply being escalated up the management tree, then their reply pings back within seconds. This says to me that they did know the answer but were just refusing to answer.The best thing is they just say yes.. or go quiet and do/say nothing.
Of course if you do escalate it and their boss is of same mindset...
But agree it is a cultural thing, also having said what written earlier, some Indian colleagues who have come to Europe have been some of the most committed, hard working, never say die colleagues I have ever had the privilege to work with.
Chim said:
honest_delboy said:
ahhhhh i seeeeeee, so after the upgrade tests were done which passed ok so 2nd node upgraded. Only under load did the problems arise.
Is this right?
From what I can gather, yes. Only mystery to me is why they had to restore from Friday. Is this right?
This looks like (from my own experience) End of day running in a bank is limited to Monday to Friday and hence upgrades are attempted at weekends, so you would only see full load of a system on the monday night when all hell would break loose... Also suggests their UAT system is not doing EOD testing with a full load of prod data and is running way under normal tolerances.
So when it breaks on Monday, they would need to restore from Fridays EOD, then run Mondays EOD and so on to catch up. Seems logical as when I designed the EOD systems for Citi and HSBC equities departments thats the standard we would do on UAT to catch up and force several EOD tests to prove the system works and we needed to do once when server overheated one night (but it didn't affect day to day business)...
I wish more people would learn what stress testing is and bleeding well test for it.
Stress testing, load testing, boundary testing, whatever the want to call it it should be done. However when I have asked for identical hardware, network, software and data setups for Test/UAT systems only to get them rejected due to cost all I can do is shrug and point to the fact I asked for proper testing and they wouldn't sign it off. Amazingly the request is quickly signed off after the first mess.
Once (years back) I was helping upgrade a banks business process production system.
So we took a copy of all the data (john smith'd it in the database) and they guys in charge of the system after getting lots of sign offs brought the data to us.
We had rented a sun server (small one, installed oracle etc) and copied all the data etc onto our system, thus replicating the system, bar the fact they were running on E10000's with huge disk arrays etc.
The upgrade (did 3 times) took 8 hours (each).. Fine, all worked should take less on the super powerful production system. This gave them confidence that our part of the upgrades (multi systems etc) would indeed work.
After 18 hours they were getting a little panicky as Monday was now starting to get closer.. 21 hours later it finished.. never worked out where the bottle neck was..
So we took a copy of all the data (john smith'd it in the database) and they guys in charge of the system after getting lots of sign offs brought the data to us.
We had rented a sun server (small one, installed oracle etc) and copied all the data etc onto our system, thus replicating the system, bar the fact they were running on E10000's with huge disk arrays etc.
The upgrade (did 3 times) took 8 hours (each).. Fine, all worked should take less on the super powerful production system. This gave them confidence that our part of the upgrades (multi systems etc) would indeed work.
After 18 hours they were getting a little panicky as Monday was now starting to get closer.. 21 hours later it finished.. never worked out where the bottle neck was..
Checked my online banking this morning, last transaction recorded was the 28th.. I had money placed in my account on the 29th so I was hopeful it would turn up tonight or perhaps tomorrow morning.. So I go to check again and get this
Ulsterbank said:
Service Temporarily Unavailable
We regret this service is temporarily unavailable and thank you for your patience.
As we continue to work hard to clear the backlog caused by our current technical issues, we have suspended our online application service until we return to normal service. We will continue to provide daily updates on our progress on our website. For the latest information please visit Help
Brilliant. So I now have to visit an ATM periodically to know if my account is even close to being up to date now.We regret this service is temporarily unavailable and thank you for your patience.
As we continue to work hard to clear the backlog caused by our current technical issues, we have suspended our online application service until we return to normal service. We will continue to provide daily updates on our progress on our website. For the latest information please visit Help
Gassing Station | News, Politics & Economics | Top of Page | What's New | My Stuff