Order posts by limited to posts

Yesterday 17:00:00
Yesterday 09:39:31
I'm sorry about this but over night our automated systems changed the balance of LNSs.

This means that this evening, once again, they will be wrong.

As such we have scheduled a ppp restart between 5pm and 6pm today to bring LNSs back to where they should be. Once again this should be a matter of a few seconds outage depending on your equipment.

What we do know is that this worked perfectly last night.

This appears to be entirely my fault as I had scheduled the rebalancing before we decided to do it manually yesterday, and did not cancel the script. I do apologise for the inconvenience.

Adrian, Director.

Started Yesterday 17:00:00
Previously expected Yesterday 18:00:00

Monday 18:17:11
Monday 11:10:11
We have been rolling out an upgrade to routers and LNSs over the weekend, and this has resulted in an unbalanced set of traffic on LNSs. This meant some customers seeing slower speeds last night, for example.

We have decided that the best move is to rebalance lines between 5pm and 6pm before the evening traffic peak.

This should represent a PPP restart and loss of service for literally a few seconds. For some customers this would happen twice.

It also means some customers will lose their graphs for today as well.

Sorry for any inconvenience, but we believe this is the best way to ensure the best service for everyone this evening.

Monday 18:06:27
The work has mostly gone to plan but is taking a tad longer than expected to complete the final stages.
Resolution Rebalancing complete - sorry for the slight overrun.
Started Monday 17:00:00
Closed Monday 18:17:11
Previously expected Monday 18:20:00

18 Nov 16:38:13
18 Nov 14:10:12
Some TalkTalk lines are missing their graphs and connection status from the Control Pages. Customers should still have normal internet access though. This is being looked in to and we expect a fix later today or this evening.
18 Nov 16:38:36
Customers should now be able to re-login and the graphs etc will show again.
Started 18 Nov 09:00:00
Closed 18 Nov 16:38:13

13 Nov 18:30:00
13 Nov 12:31:38
Our constant quality monitoring has identified severe packet loss and latency on some TalkTalk lines connected through the South west of England and South West Wales. An incident has been raised into TalkTalk
13 Nov 12:35:01
Here is an example graph
13 Nov 12:49:42
TalkTalk are aware, it looks like a fibre break in the Plymouth area.
13 Nov 12:53:39
TalkTalk are treating this as priority one incident. More info to follow as and when.
13 Nov 16:56:08
Update from TT: "The incident is still on going and we expect it to be back up by 10pm"
13 Nov 19:00:51
lines are looking back to normal as of 18:30
Started 13 Nov 09:29:25
Closed 13 Nov 18:30:00

9 Nov 10:30:59
27 Apr 14:56:58
Previously Openreach had advised that they intend to run a trial starting today on a small set of lines; where devices aren’t retransmission compatible in the upstream to see if changing certain parameters removes the latency and maintains the other benefits of retransmission. They have now advised us that the trial start date has been put back by two weeks (no idea why)

So if you have an FTTC line that is affected by this then please drop an email to support and we can include it in the list of affected lines that we will get included in the trial.

11 May 11:07:12
Openreach have advised they will start loading the new DLM profiles to lines on Tuesday morning as part of regular DLM runs. Customers that are on the trial will notice a loss of sync when the new profiles are updated.
3 Jun 10:36:31
Apologise for the delay in updating this post. BT have confirmed that all trial lines have been loaded with the new profiles, further to this BT have confirmed that all other affected lines have now had the new profiles loaded. That is all lines across all providers.
Started 27 Apr 14:51:39

15 Oct 16:50:47
15 Oct 16:50:47
We are running a trial, open to everyone on a TalkTalk connected ADSL line (non-annex M at the moment), which will allow us to set the line profile to 3dB. Typically this will give a speed improvement on short lines. Email in to trials@aa.net.uk if you're interested. It's easy for us to move the line back if it is unstable, there is no cost difference.
Started 15 Oct 15:03:03

13 Oct 16:48:23
13 Oct 15:52:41
We had a network incident starting from just after 15:30 today lasting about 15 minutes. This looks like a DDOS attack but we are still investigating.
13 Oct 15:53:35
Normal service resumed at around 15:41
Started 13 Oct 15:30:00
Closed 13 Oct 16:48:23

22 Sep 11:35:00
21 Sep 13:26:39
There is some sort of routing problem affecting customers. We are still investigating
21 Sep 13:35:40
Routing/traffic is back to normal, we continue to monitor...
22 Sep 10:24:49
The burst of traffic that caused this disruption has not returned. We do apologise for the inconvenience this caused.
Started 21 Sep 13:20:00
Closed 22 Sep 11:35:00

10 Sep 10:05:21
25 May 22:24:55
We're seeing peak time (evening) congestion on lines at BERMONDSEY exchange again. It started on evening of 29th April. We've reported this previously, on 14 Jan and was fixed then on the 15 Jan. We'll update this post shortly.
2 Jun 13:51:26
No updates as yet. We are chasing TT again today.
5 Jun 09:18:13
Sadly no update, we'll chase this via alternate channels!
10 Sep 10:05:34
It appears that the issue is resolved.
Started 25 May 22:23:31
Closed 10 Sep 10:05:21

10 Sep 09:54:03
4 May 21:41:59
We are are seeing congestion on HORNDEAN and WATERLOOVILLE exchanges (Hampshire). This is usually noticeable in the evenings. This will be reported to BT, and we'll update this post with updates.
10 Sep 09:54:17
Congestion issues appear to be resolved at both locations.
Started 4 May 21:38:34
Closed 10 Sep 09:54:03

10 Sep 09:53:10
19 Jun 11:28:43
We are seeing around 1-2% packet loss on the MAIDA VALE Exchange. This has been reported to the TSO team within BT Wholesale.
10 Sep 09:53:19
Packet loss appears to be gone
Broadband Users Affected 0.50%
Started 19 Jun 11:26:46
Closed 10 Sep 09:53:10
Previously expected 21 Jun 13:00:00

10 Sep 09:50:07
11 Aug 11:35:34
We have noticed congestion one some of our TalkTalk circuits. customers affected are seeing an increase of latency and some packet loss. We have reported this to TalkTalk and as soon as we have an update we will update this post.
11 Aug 11:42:09
Here is an example graph
11 Aug 14:49:09
Update from TT We lost multiple 10G circuits to our transmission network due to a BT fibre break in the Ipswich area - Openreach have applied for a MBORC on the issue. There was congestion in the Ipswich and Cambridgeshire wider area as a result which was resolved shortly after midnight.
Started 10 Aug 17:32:43
Closed 10 Sep 09:50:07

10 Sep 09:49:30
14 Aug 11:50:09
We have been seeing increased packet loss and latency on certain line on the Forest hill exchange. It appears one VLAN is currently running hot. BT are aware and are investigating.
18 Aug 10:31:33
BT believe the SVLAN has been upgraded. We will monitor the circuits tonight.
10 Sep 09:49:48
Issues appear to be resolved.
Started 14 Aug 11:46:37 by AA Staff
Closed 10 Sep 09:49:30
Previously expected 17 Aug 09:00:00

26 Aug 23:30:00
26 Aug 00:16:15
We've seen a couple of incidents this evening where customers have seen packet loss and some DSL sessions have blipped. It seems to have been caused by a traffic flood, and we're investigating the cause. This status is posted as Minor as things look OK at the moment.
26 Aug 20:33:37
We've had another set of blips affecting internet access for some customers this evening (20:15ish). We are looking in to the cause of this.
26 Aug 21:07:47
This is problem is continuing to cause disruption, please bear with us as we work to get to the bottom of this.
26 Aug 21:37:02
Traffic is has been normal for the past 30 mins or so, we've not yet tracked down the cause and are still looking on to this.
26 Aug 22:37:34
We believe we have found the cause of the network issues. More details to follow in the morning.
26 Aug 23:05:49
Unfortunately we are still having problems, we're still working on this.
26 Aug 23:33:38
We've done further work on this, traffic is back to normal at the moment and we are monitoring the situation.
Resolution The intermittent problems on the evenings of 25th and 26th were caused by a distributed denial of service attack against one of our customers. This type of attack does happen from time to time and we have systems which usually stop an attack in its tracks very quickly. However, in this case the attack was not automatically blocked. Due to the nature of the attack it did take longer than we would have liked to pin point the target of the attack, and this was eventually done at around 23:30 on the 26th. Moving on, we will be looking in to why this was not automatically blocked and how we can improve other systems with a view to prevent this from being a problem in the future. We do apologise for the inconvenience this caused.
Closed 26 Aug 23:30:00

27 Jul 10:36:06
16 Jun 11:32:55
As some of you may know, the new system for migrations of broadband lines comes in this weekend. See http://aa.net.uk/news-20150601-not.html

We expect to enable the new notice emails during this week, maybe even today. These will be sent for customers migrating away from us, and migrating to us (when order commits) and include the options for cancelling the migration.

We are also taking this opportunity to include notices for services ceasing - either because you have requested a cease, or unsolicited (e.g. if ceasing phone line causes cease of broadband).

In any case, if leaving us, the notice details any early termination charges.

If any issues, concerns, or questions, please let us know.

16 Jun 18:24:44
We are now sending NoT emails. Please let us know any issues.
Started 16 Jun
Previously expected 20 Jun

9 Jul 17:15:00
9 Jul 16:35:04
We are reloading one of our route reflectors - this should have little of no impact as routing should automatically fall over to the other.

If this does not resolve the issues we may have to do more work.

Started 9 Jul 17:00:00
Closed 9 Jul 17:15:00
Previously expected 9 Jul 17:10:00

1 Jul 04:34:32
1 Jul 03:11:16
From 2:18 today, we are seeing packet loss and some lines dropping their connection.
Resolution We've identified the caused of this, and traffic is back to normal.
Started 1 Jul 02:18:00
Closed 1 Jul 04:34:32

24 Jun 16:27:39
24 Jun 16:27:39
When Home::1 reaches quota you can opt to have the service stop, slow down, or auto top-up.

Slow mode was around 250kb/s. However, we have changed this so that the speed depends on the tariff. For 100GB users the slow mode will now be around 330kb/s and for 200GB users it will be around 660kb/s. The idea is that at that speed you could not use the whole of the coming month's quote even flat out.

Started 24 Jun 16:25:00

19 Jun 16:28:46
31 May 15:28:31
With the up-coming changes to broadband migrations and the abolishing of Migration Authorisation Codes (MACs), AAISP has launched a new "anti-slamming" service to allow customers to "lock" their line against unwanted migrations to another provider.

This extra service, which is completely free, works on much the same principle as "domain locking" where domains can be locked against migration.

The process is simple and allows each line to be locked against migrations for broadband or the underlying copper pair "phone line" part of the service. It is just a "standing order" from the customer to AAISP to reject all migrations. From the 20th, any migration request that then comes in is automatically rejected if anti-slamming is enabled. An email is sent to advise the customer of what happened, including a simple link to turn off the anti-slamming if they do wish to migrate after all.

The anti-slamming service has been provided by popular demand after many customers expressed concerns that their lines could be "slammed" (maliciously taken over by other telcos) or that mistakes could lead to unwanted migrations, and it would be very easy to miss the notice of transfer that is sent before the migration goes ahead. OFCOM do not seem to have created any "fast correct" of mistakes, so an unwanted migration could mean waiting another 10 working days to fix the situation.

Slamming is just one of the many concerns over the new migration process. There may still be ways LLU providers can take over lines, as can happen now without a MAC, but this new service should avoid mistakes and give customers peace of mind.

The control pages for the line include a simple link to enable or disable the anti-slamming service.

Started 31 May 15:27:25
Previously expected 20 Jun

19 Jun 16:28:30
14 May 15:42:00

We are please to inform customers that we are changing the entry-level router that we supply to our customers. From next week we'll be shipping the ZyXEL VMG1312 by default instead of the Technicolor.

Since around 2012 we have been providing the Technicolor TG852, which was the first consumer level router to support IPv6. With the advent of wires-only FTTC and the need for a more flexible and easy to use router we have been looking for a replacement. The ZyXEL VMG1312 is able to do ADSL, FTTC and PPPoE and is flexible enough to be used on most of our lines. We have been working with ZyXEL over the past few months to iron out bugs that we have found. There are still some bugs to be fixed and these are detailed on our Support site. The biggest bug is the lack of 1500 byte MTU when running in bridge mode, however, ZyXEL expect to have this fixed soon and in the meantime FTTC installations will be installed with the Openreach modem.

More information about the router: https://support.aa.net.uk/Category:ZyXEL_VMG1312

Started 14 May 13:12:00

19 Jun 16:27:57
19 Jun 16:27:57
You no longer need to obtain a Migration Authorisation Code to migrate broadband. Simple contact the new provider with your details to arrange the migration.

If you wish to migrate to us, please complete the order, and you will not be asked for a MAC.

If you wish to migrate away, simply contact the new provider (from tomorrow) and they can arrange the migration.

The lead time has increased from 5 working days to 10 - a significant added delay for which you can thank OFCOM.

Started 19 Jun 16:25:59
Previously expected 20 Jun

18 Jun 15:00:00
[Broadband] - BT Blip - Closed
18 Jun 14:50:08
A number of BT lines dropped out at about 14:44, this appears to be an issue on the BT side, all lines have now recovered. We are chasing Bt for an explanation.
18 Jun 15:03:46
Some lines are still blipping, we are investigating further.
Resolution The blip seems to have gone away - we may move some lines that ended up on the wrong LNS back later. This is not the first time this link to BT had blipped and we are chasing with BT.
Started 18 Jun 14:47:59
Closed 18 Jun 15:00:00

16 Jun 18:25:11
13 Feb 14:29:06
We currently supply the Technicolor TG582 for most ADSL services, but we are considering switching to a new router, the ZyXEL VMG1312-B

It is very comprehensive and does both ADSL and VDSL as well as bridging and wifi. It means we can have one router for all service types. As some of you may know, BT will be changing FTTC to be "wires only" next year, and so a VDSL router will be needed.

We have a small number available now for people to trial - we want to test the routers, our "standard config" and the provisioning process.

Please contact trial@aa.net.uk or #trial on the irc server for more information.

P.S. Obviously it does Internet Protocol, the current one, IPv6 and the old one IPv4

Obviously this initial trial is limited number of routers which we are sending out at no charge to try different scenarios. However, we expect to be shipping these as standard later in the month, and they will be available to purchase on the web site.

13 Feb 15:49:08
Thanks for all the emails and IRC messages about trialling the new routers. We will contact customers back on Monday to arrange shipping of the units.
16 Feb 10:43:54
We now have enough trialists for the new router, we will contact a selection of customers today to arrange delivery of the routers. Thanks
Started 13 Feb 14:25:34

16 Jun 18:24:58
3 Jun 11:30:45

We are now processing FTTC orders as 'wires only'. This means:

  • An Openreach engineer appointment is not required if the phone line already exists
  • Openreach will not provide a modem.
This makes the installation easier to arrange as no onsite appointment is required. There are a number of FTTC (VDSL) modems and routers available on the market, we are providing the ZyXEL VMG1312 router which can act as an all-in-one modem/router or just as a modem in a similar fashion to the original Openreach modem.

Do contact us for more information

Started 3 Jun 11:00:00

10 Jun 02:21:49
10 Jun 10:26:47
It seems that one of our BT links had an issue over night. It was only one of the four links we have and only BT services that were affected. Lines reconnected. If you have not reconnected, try power cycling your router. We don't have any explanation from BT as yet.
Started 10 Jun 02:03:19
Closed 10 Jun 02:21:49

1 Jun 23:30:51
1 Jun 23:11:35
We're seeing odd routing blips across our network. We're investigating.
1 Jun 23:31:30
Traffic is back to normal. We'll investigate this further and will update this post in the morning.
Resolution This looked like it was some sort of DOS that affected our routing, possibly other ISPs too.
Started 1 Jun 22:55:00
Closed 1 Jun 23:30:51

12 May 10:27:29
5 May 10:21:17
BT have had trouble with this exchange (http://aastatus.net/2052) but we are now seeing evening congestion on TalkTalk connected lines. We have reported this and will update this post accordingly.
5 May 12:38:47
TalkTalk have fixed a misconfiguration at their end. This should now be resolved. We'll check again tomorrow.
7 May 08:55:48
Lines still seem to be congested, this is being looked in to by TalkTalk.
13 May 12:13:36
Update from TT 'This should now properly be resolved. A faulty interface meant that we were running at 2/3 of capacity - only a few of your subscribers would have suffered last night hopefully but should now all be good.'
Started 5 May 10:00:00
Closed 12 May 10:27:29

2 Jun 11:18:33
2 Jun 11:15:34
At 11:07 one of our links to BT blipped - this caused a number of BT DSL lines to drop. Lines are reconnecting at the moment.
2 Jun 11:18:56
Most lines are now back online.
Started 2 Jun 11:07:00
Closed 2 Jun 11:18:33

27 May 14:48:58
27 May 10:37:50
One of our routers is misbehaving - the one we planned to reload shortly as it happens, and this seems to be impacting mobile data SIMs.
27 May 10:46:58
This appears to be a LINX issue
27 May 10:53:20
We have managed to shut down the port to LINX while this is resolved.
27 May 11:01:41
This looks to be impacting lots of ISPs, but traffic will go via alternative peering and transit whilst LINX have issues and so this should have minimal impact on customers at present.
27 May 12:09:13
LINX are clearly having some major issues on one of the LANs, and we remain disconnected until this is resolved. Once it is resolved, we may reload one of our routers so as to add more logging and allow us to more carefully monitor further issues.
27 May 14:39:14
LINX have rectified the issue, so we will be re-enabling LINX ports shortly.
Resolution All working now.
Started 27 May 10:30:00
Closed 27 May 14:48:58

27 May 10:53:48
27 May 10:12:22
We have been encountering issues with LINX which result in some routes being "Black-holed" (i.e. going nowhere).

The system normally handles this without any problem, identifying an infeasible route and removing from the routing table to allow alternative routes a chance. However we have identified an edge case where this does not happen.

We plan to do a router upgrade today, as this issue is causing customers problems at present. This is one router and should be pretty seamless in that routing should go via other routers whilst it happens.

However, there is a small risk of some routing blips during the process.

Resolution Upgrade completed
Started 27 May 10:00:00
Closed 27 May 10:53:48
Previously expected 27 May 17:00:00

25 May 21:30:00
25 May 21:45:20
From around 6:30pm there was IPv6 routing problems to some hosts via our Linx peering - notably to Google.
Resolution A temporary workaround was applied at 9:30pm and routing has been restored. Staff were alerted to this via customers using the 'MSO' SMS facility, however, do to individual staff circumstances it was not until 9pm that staff were able to respond. We do apologise for the time this took, it is very rare for all the staff that are alerted in this way to all be unavailable at the same time. We'll consider what can be done to improve this.
Started 25 May 18:30:00
Closed 25 May 21:30:00

13 May 12:16:45
26 Mar 09:53:31

Over the past couple of weeks we have seen FTTC lines drop and reconnect with in increase in latency of around 15ms. This is seen on the monitoring graphs as a thicker blue line.

Upon first glance it looks as if interleaving has been enabled, but a line test shows that this is not the case.

We've been in contact with BT and it does look like BT are rolling out a new profile on to their Huawei DSLAMs in the local green cabinets. It has been expected that BT would be rolling out this new profile, but we didn't expect such an increase in latency.

The profile adds 'Physical Retransmission (ReTX) technology (G.INP / ITU G.998.4)' which helps with spikes of electromagnetic interference and can make lines more stable.

We would hope to have control over the enabling and disabling of this profile, but we don't. Line profiles with FTTC is managed by BT Openreach and are tricky for us and even BT Wholesale to get adjusted.

We're still discussing this with BT and will update this post with news as we have it.

26 Mar 09:55:29
26 Mar 10:48:37
This has been escalated to the head of fibre deployment within BT Wholesale and we are expecting an update by the end of the day.
26 Mar 11:12:08
Further information about G.INP:
  • http://www.ispreview.co.uk/index.php/2015/01/bt-enables-physical-retransmission-g-inp-fttc-broadband-lines.html
  • http://www.thinkbroadband.com/news/6789-impulse-noise-protection-rolling-out-on-openreach-vdsl2.html
  • http://forum.kitz.co.uk/index.php?topic=15099.0
...among others.
27 Mar 16:46:22
BT have asked us for further information which we have provided to them, we don't expect an update now until Monday
9 Apr 14:26:19
This is still ongoing with BT Wholesale and BT Openreach.
16 Apr 15:58:53
This has been escalated to a very senior level within BT and we are expecting a proper update in the next few days.
24 Apr 13:01:45
We have just received the below back from BT on this

Following communications from a small number of BTWholesale FTTC comms providers regarding Openreach’s implementation of Retransmission and the identification of some your customers who have seen increased latency on some lines for some applications since retransmission was applied. Over the last 4 weeks I have been pushing Openreach to investigate,

feedback and provide answers and options related to this issue. As a result attached is a copy of a briefing from Openreach; sent to their CPs today, on how RetX works and what may have caused this increase latency.

This info is being briefed to all BTWholesale customers via our briefing on Saturday morning 25/4/15 but as you have contacted me direct I’m sending this direct as well as providing ann opportunity to participate in a trial.

Openreach have also advise me this afternoon that they intend to run a trial next week (w/c 25/4/15) on a small set of lines; where devices aren’t retransmission compatible in the upstream to see if changing certain parameters removes the latency and maintains the other benefits of retransmission. The exact date lines will be trialled has yet to be confirmed.

However, they have asked if I have any end users who would like to be included in this trail. To that end if you have particular lines you’d like to participate in this trial please can you provide the DN for the service by 17:00 on Monday 28th April so I can get them included.

This is a trial of a solution and should improve latency performance but there is a risk that there may be changes to the headline rate.

5 May 22:28:25
Update to Trial here: https://aastatus.net/2127
Started 26 Mar 09:00:00
Closed 13 May 12:16:45

12 May 13:41:47
11 May 18:07:36
We have changed the way we apply minimum terms.

Instead of charging for all of the service to the end of the minimum term, we now charge an early termination fee for the period from the cease/migrate to the end of the term. This is a simple fee based on the tariff and line type. We are also scrapping the 30 day notice requirement.

Whilst the old system was simple, it did not fit OFCOM rules for the new migration system. We think the new system is equally simple, and saves customers money. As such the change has been introduced today.

More details here http://aa.net.uk/news-20150511-minterm.html

Started 11 May 18:00:00

9 May 14:31:50
5 May 22:22:42

TalkTalk is performing essential maintenance on its internal infrastructure. This work will happen in the early hours of the 8th May 2015. This work will mean that services may be lost for up to 1 hour between midnight and 6am. This is likely to affect the following exchanges:

  • Yatton
  • Didcot
  • Cheltenham
  • Swindon
  • Blunsdon
  • Cirencester
  • BOX
  • Chippenham
  • Clevedon
  • PILL
  • Brimscombe
  • Dursley
  • Calne
  • Devizes
  • Melksham
  • Worle
  • Weston Super Mare
  • Downend
  • Midsomer Norton
  • Radstock
  • West
  • Frome
  • Trowbridge
  • Westbury
  • Portishead
  • Kingswood
  • Kingsmead
  • Bristol North
  • South
  • Glastonbury
  • Wells
We apologise for any inconvenience that these works may cause you
Started 8 May by TalkTalk
Closed 9 May 14:31:50

23 Apr 2014 10:21:03
01 Nov 2013 15:05:00
We have identified an issue that appears to be affecting some customers with FTTC modems. The issue is stupidly complex, and we are still trying to pin down the exact details. The symptoms appear to be that some packets are not passing correctly, some of the time.

Unfortunately one of the types of packet that refuses to pass correctly are FireBrick FB105 tunnel packets. This means customers relying on FB105 tunnels over FTTC are seeing issues.

The work around is to remove the ethernet lead to the modem and then reconnect it. This seems to fix the issue, at least until the next PPP restart. If you have remote access to a FireBrick, e.g. via WAN IP, and need to do this you can change the Ethernet port settings to force it to re-negotiate, and this has the same effect - this only works if directly connected to the FTTC modem as the fix does need the modem Ethernet to restart.

We are asking BT about this, and we are currently assuming this is a firmware issue on the BT FTTC modems.

We have confirmed that modems re-flashed with non-BT firmware do not have the same problem, though we don't usually recommend doing this as it is a BT modem and part of the service.

04 Nov 2013 16:52:49
We have been working on getting more specific information regarding this, we hope to post an update tomorrow.
05 Nov 2013 09:34:14
We have reproduced this problem by sending UDP packets using 'Scapy'. We are doing further testing today, and hope to write up a more detailed report about what we are seeing and what we have tested.
05 Nov 2013 14:27:26
We have some quite good demonstrations of the problem now, and it looks like it will mess up most VPNs based on UDP. We can show how a whole range of UDP ports can be blacklisted by the modem somehow on the next PPP restart. It is crazy. We hope to post a little video of our testing shortly.
05 Nov 2013 15:08:16
Here is an update/overview of the situation. (from http://revk.www.me.uk/2013/11/bt-huawei-fttc-modem-bug-breaking-vpns.html )

We have confirmed that the latest code in the BT FTTC modems appears to have a serious bug that is affecting almost anyone running any sort of VPN over FTTC.

Existing modems seem to be upgrading, presumably due to a roll out of new code in BT. An older modem that has not been on-line a while is fine. A re-flashed modem with non-BT firmware is fine. A working modem on the line for a while suddenly stopped working, presumably upgraded.

The bug appears to be that the modem manages to "blacklist" some UDP packets after a PPP restart.

If we send a number of UDP packets, using various UDP ports, then cause PPP to drop and reconnect, we then find that around 254 combinations of UDP IP/ports are now blacklisted. I.e. they no longer get sent on the line. Other packets are fine.

Sending 500 different packets, around 254 of them will not work again after the PPP restart. It is not actually the first or last 254 packets, some in the middle, but it seems to be 254 combinations. They work as much as you like before the PPP restart, and then never work after it.

We can send a batch of packets, wait 5 minutes, PPP restart, and still find that packets are now blacklisted. We have tried a wide range of ports, high and low, different src and dst ports, and so on - they are all affected.

The only way to "fix" it, is to disconnect the Ethernet port on the modem and reconnect. This does not even have to be long enough to drop PPP. Then it is fine until the next PPP restart. And yes, we have been running a load of scripts to systematically test this and reproduce the fault.

The problem is that a lot of VPNs use UDP and use the same set of ports for all of the packets, so if that combination is blacklisted by the modem the VPN stops after a PPP restart. The only way to fix it is manual intervention.

The modem is meant to be an Ethernet bridge. It should not know anything about PPP restarting or UDP packets and ports. It makes no sense that it would do this. We have tested swapping working and broken modems back and forth. We have tested with a variety of different equipment doing PPPoE and IP behind the modem.

BT are working on this, but it is a serious concern that this is being rolled out.
12 Nov 2013 10:20:18
Work on this in still ongoing... We have tested this on a standard BT retail FTTC 'Infinity' line, and the problem cannot be reproduced. We suspect this is because when the PPP re-establishes a different IP address is allocated each time, and whatever is session tracking does not match the new connection.
12 Nov 2013 11:08:17

Here is an update with some a more specific explanation as to what the problem we are seeing is:

On WBC FTTC, we can send a UDP packet inside the PPP and then drop the PPP a few seconds later. After the PPP re-establishes, UDP packets with the same source and destination IP and ports won't pass; they do not reach the LNS at the ISP.

Further to that, it's not just one src+dst IP and port tuple which is affected. We can send 254 UDP packets using different src+dest ports before we drop the PPP. After it comes back up, all 254 port combinations will fail. It is worth noting here that this cannot be reproduced on an FTTC service which allocates a dynamic IP which changes each time PPP re-established.

If we send more than 254 packets, only 254 will be broken and the others will work. It's not always the first 254 or last 254, the broken ones move around between tests.

So it sounds like the modem (or, less likely, something in the cab or exchange) is creating state table entries for packets it is passing which tie them to a particular PPP session, and then failing to flush the table when the PPP goes down.

This is a little crazy in the first place. It's a modem. It shouldn't even be aware that it's passing PPPoE frames, let along looking inside them to see that they are UDP.

This only happens when using an Openreach Huawei HG612 modem that we suspect has been recently remotely and automatically upgraded by Openreach in the past couple of months. Further - a HG612 modem with the 'unlocked' firmware does not have this problem. A HG612 modem that has probably not been automatically/remotely upgraded does not have this problem.

Side note: One theory is that the brokenness is actually happening in the street cab and not the modem. And that the new firmware in the modem which is triggering it has enabled 'link-state forwarding' on the modem's Ethernet interface.

27 Nov 2013 10:09:42
This post has been a little quiet, but we are still working with BT/Openreach regarding this issue. We hope to have some more information to post in the next day or two.
27 Nov 2013 10:10:13
We have also had reports from someone outside of AAISP reproducing this problem.
27 Nov 2013 14:19:19
We have spent the morning with some nice chaps from Openreach and Huawei. We have demonstrated the problem and they were able to do traffic captures at various points on their side. Huawei HQ can now reproduce the problem and will investigate the problem further.
28 Nov 2013 10:39:36
Adrian has posted about this on his blog: http://revk.www.me.uk/2013/11/bt-huawei-working-with-us.html
13 Jan 2014 14:09:08
We are still chasing this with BT.
03 Apr 2014 15:47:59
We have seen this affect SIP registrations (which use 5060 as the source and target)... Customers can contact us and we'll arrange a modem swap.
23 Apr 2014 10:21:03
BT are in the process of testing an updated firmware for the modems with customers. Any customers affected by this can contact us and we can arrange a new modem to be sent out.
7 May 22:56:52
Just a side note on this, we're seeing the same problem on the ZyXEL VMG1312 router which we are teting out and which uses the same chipset: info and updates here: https://support.aa.net.uk/VMG1312-Trial
Resolution BT are testing a fix in the lab and will deploy in due course, but this could take months. However, if any customers are adversely affected by this bug, please let us know and we can arrange for BT to send a replacement ECI modem instead of the Huawei modem. Thank you all for your patience.

BT do have a new firmware that they are rolling out to the modems. So far it does seem to have fixed the fault and we have not heard of any other issues as of yet. If you do still have the issue, please reboot your modem, if the problem remains, please contact support@aa.net.uk and we will try and get the firmware rolled out to you.
Started 25 Oct 2013
Closed 23 Apr 2014 10:21:03

7 May 09:53:49
09 Dec 2014 11:20:04
Some lines on the LOWER HOLLOWAY exchange are experiencing peak time packet loss. We have reported this to BT and they are investigating the issue.
11 Dec 2014 10:46:42
BT have passed this to TSO for investigation. We are waiting for a further update.
12 Dec 2014 14:23:56
BT's Tso are currently investigating the issue.
16 Dec 2014 12:07:31
Other ISPs are seeing the same problem. The BT Capacity team are now looking in to this.
17 Dec 2014 16:21:04
No update to report yet, we're still chasing BT...
18 Dec 2014 11:09:46
The latest update from this morning is: "The BT capacity team have investigated and confirmed that the port is not being over utilized, tech services have been engaged and are currently investigating from their side."
19 Dec 2014 15:47:47
BT are looking to move our affected circuits on to other ports.
13 Jan 10:28:52
This is being escalated further with BT now, update to follow
19 Jan 12:04:34
This has been raised as a new reference as the old one was closed. Update due by tomorrow AM
20 Jan 12:07:53
BT will be checking this further this evening so we should have more of an update by tomorrow morning
22 Jan 09:44:47
An update is due by the end of the day
22 Jan 16:02:24
This has been escalated further with BT, update probably tomorrow now
23 Jan 09:31:23
we are still waiting for a PEW to be relayed to us. BT will be chasing this for us later on in the day.
26 Jan 09:46:03
BT are doing a 'test move' this evening where they will be moving a line onto another VLAN to see if that helps with the load, if that works then they will move the other affected lines onto this VLAN. Probably Wednesday night.
26 Jan 10:37:45
there will be an SVLAN migration to resolve this issue on Wednesday 28th Jan.
30 Jan 09:33:57
Network rearrangement is happening on Sunday so we will check again on Monday
2 Feb 14:23:12
Network rearrangement was done at 2AM this morning, we will check for paclet loss and report back tomorrow.
3 Feb 09:46:49
We are still seeing loss on a few lines - I am not at all happy that BT have not yet resolved this. A further escalation has been raised with BT and an update will follow shortly.
4 Feb 10:39:03
Escalated futher with an update due at lunch time
11 Feb 14:14:58
We are getting extremly irritated with BT on this one, it should not take this long to add extra capaity in the affected area. Rocket on it's way to them now ......
24 Feb 12:59:54
escalated further with BT, update due by the end of the day.
2 Mar 09:57:59
We only have a few customers left showing peak time packet loss and for now the fix will be to move them onto another MSAN, I am hoping this will be done in the next few days. We really have been pushing BT hard on this and other areas where we are seeing congestion. I am please that there are now only a handful of affected customers left.
17 Mar 11:21:33
We have just put a boot up BT on this, update to follow.
2 Apr 13:16:10
BT have still not fixed the fault so we have moved some of the affected circuits over to TalkTalk and I am pleased to say that we are not seieng loss on those lines. 100% this is a BT issue and I am struggling to understand why they have still not tracked the fault down.
Closed 7 May 09:53:49
Previously expected 1 Feb 09:34:04 (Last Estimated Resolution Time from AAISP)

7 May 09:52:18
11 Mar 11:39:17
We are seeing some evening time congestion on all BT 21CN lines that connect through BRAS's 21CN-BRAS-RED1-MR-DH up to 21CN-BRAS-RED13-MR-DH I suspect one of the BT nodes is hitting limits some evenings as we don't see the higher latency every night. This has been reported into BT and we will update this past as soon as they respond back.
11 Mar 11:44:06
Here is an example graph
12 Mar 12:00:45
This has been escalated further to the BT networK guys and we can expect an update within the next few hours.
17 Mar 15:41:18
Work was done on this overnight so I will check again tomorrow morming and post another update.
18 Mar 11:38:25
The changes BT made over night have made a significant difference to the latency, still seeing it slightly higher than we would like so we will go back to then again.
19 Mar 14:54:44
Unfortunately the latency has increased again so whatever BT did two nights ago has not really helped. We are chasing again now.
23 Mar 14:07:53
BT have still not pinpointed the issue so it has been escalated further.
27 Mar 13:03:38
Latency is hardly noticeable now but we are still chasing BT on sorting the actual issue, update will be mOnday now.
30 Mar 10:04:14
BT have advised that they are aware of the congestion issue at Manchester, and the solution they have in place is to install some additional edge routers, they are already escalating on this to bring the date in early, currently the date is May. Obviously May is just not acceptable and we are doing all we can to get BT to bring this date forward.
2 Apr 12:28:39
We have requested a further escalation within BT, the time scales they have given for a fix is just not acceptable.
13 Apr 15:12:23
The last update from BT was 'is latency issue has been escalated to high level. BT TSO are currently working on a resolution and are hoping to move into the testing phase soon. We will keep you updated as we get more information' I am chasing for another update now.
16 Apr 16:01:15
We are still chasing BT up on bringing the 'fix' forward. Hopefully we will have another response by the morning.
21 Apr 13:25:21
The latest update from BT: We have identified a solution to the capacity issue identified and are looking to put in a solution this Friday night...
24 Apr 15:25:51
BT have added more capacity on tehir network and last night the latency looked fine. We will review this again on Monday.
Started 11 Mar 01:35:37
Closed 7 May 09:52:18

29 Apr 15:23:27
29 Apr 14:43:36
A third of our BT lines bliped - this looks to be an issue with routing on one of our LNSs in to BT.
29 Apr 14:50:18
Many lines are failing to reconnect properly, we are investigating this.
29 Apr 14:57:42
Lines are connecting successfully now
29 Apr 15:23:27
The bulk of lines are back onlne. There are a small number of lines that are still failing to reconnect. These are being looked in to.
29 Apr 15:36:54
The remain lines are reconnecting successfully now.
Resolution I wanted to try and explain more about what happened today, but it is kind of tricky without saying "Something crazy in the routing to/from BT".

We did, in fact make a change - something was not working with our test LNS and a customer needed to connect. We spotted that, for some unknown reason, the routing used a static route internally instead of one announced by BGP, for just one of the four LNSs, and that on top of that the static route was wrong, hence the test LNS not working via that LNS. It made no sense, and as all three other LNSs were configured sensibly we changed the "A" LNS to be the same, after all, this is clearly a config that just worked and was no problem, or so it seemed.

Things went flappy, but we could not see why. It looks like BGP in to BT was flapping, so people connected and disconnected rather a lot. We returned the config and things seemed to be fixed for most people, but not quite all. This made no sense. Some people are connecting and going on line, and then falling off line.

The "fix" to that was to change the endpoint LNS IP address used by BT to an alias on the same LNS. We have done this in the past where BT have had a faulty link in a LAG. We wonder if this issue was "lurking" and the problem we created showed it up. This shows that there was definitely an issue in BT somehow as the fix should not have made any difference otherwise.

What is extra special is that this looks like it has happened before - the logs suggest the bodge of a static route was set up in 2008, and I have this vague recollection of a mystery flappiness like this which was never solved.

Obviously I do apologise for this, and having corrected the out of data static route this should not need touching again, but damn strange.

Started 29 Apr 14:38:00
Closed 29 Apr 15:23:27
Previously expected 29 Apr 14:50:00

25 Apr 18:46:00
25 Apr 18:48:19
There was an unexpected blip in routing - we are looking in to it.
Started 25 Apr 18:44:00
Closed 25 Apr 18:46:00
Previously expected 25 Apr 22:46:00

20 Apr 09:54:30
20 Apr 09:54:30
Customers will have received an email from us. Apologies for not PGP signing it. It asks you to go to a secure link on our control pages and confirm (one click) that you consent to receive notices via email.

Yes, I know it is crazy, and it is already part of our terms, and you already know we email notices, and that this email is a notice we have emailed you... Sorry but OFCOM insist we get *explicit* consent to send some notices we send.

We'd appreciate it if you just click the link and then the confirm button.

We'll email you again if you don't, sorry. If you are not happy about this, please do complain to OFCOM. Thank you.

21 Apr 18:23:01
We have resent the email to all of those that have not followed the link and confirmed. This time, PGP signed. Sorry for any concern the previous email caused.
21 Apr 18:29:59
I'd also like to thank the *thousands* of people that have confirmed their consent so far.
Started 19 Apr
Previously expected 1 Jun

17 Apr 15:54:16
15 Apr 13:16:15
Some customers on the Bradwell Abbey exchange are currently experiencing an outage. We have received reports from FTTP customers, however this may also affect customers using other services. BT have advised that they are currently awaiting a delivery for a new card at this exchange. We will chase BT for updates and provide them as we receive them.
15 Apr 15:47:41
I have requested a further update from BT.
16 Apr 08:07:15
Openreach AOC and PTO are investigating further at this time. We will reach out for an update later today.
16 Apr 10:32:55
BT have advised that a Cable down is the root cause at this time.
16 Apr 15:51:50
PTO are still onsite. I have asked for an ECD, however OpenReach are not supplying that information, due to being fibre work.
17 Apr 10:19:33
OpenReach have stated, they are hoping for a completion on the fibre today and resource is being tasked out. OpenReach have stated this is only an estimate and not set in stone.
17 Apr 14:49:46
Some customers are reporting a restored service. BT advise that teams are still on site to resolve this P1 issue.
17 Apr 15:55:25
The cable down issue affecting customers using the Bradwell Abbey exchange has now been resolved.
Started 15 Apr 12:55:00 by AAISP Staff
Closed 17 Apr 15:54:16
Cause BT

16 Apr 16:00:24
27 Mar 14:03:52
We are seeing packet loss on all lines connected through 21cn-BRAS-RED8-SL the loss is all through the day/night started 10:08 on the 25th. This has been reported to BT
27 Mar 14:07:22
Here is an example graph:
30 Mar 14:37:04
BT claimed to have fixed this but our monotoring is still seeing the loss, BT chased further
Broadband Users Affected 0.01%
Closed 16 Apr 16:00:24

16 Apr 15:59:33
2 Feb 10:10:46
We are seeing low level packet loss on BT lines connected to the Wapping exchange - approx 6pm to 11pm every night. Reported to BT...
2 Feb 10:13:57
Here is an example graph:
3 Feb 15:55:40
Thsi has been escalated further with BT
4 Feb 10:27:37
Escalated further with BT, update due after lunch
11 Feb 14:18:00
Still not fixed, we are arming yet another rocket to fire at BT smiley
24 Feb 12:58:51
escalated further with BT, update due by the end of the day.
2 Mar 10:00:11
Again the last few users seeing packet loss will be moved onto another MSAN in the next few days.
12 Mar 12:02:57
Updatew expected in the next few hours
17 Mar 11:19:48
A further escalation has been raised on this, update by the end of the day
30 Mar 15:35:32
This has been escalated to the next level
Broadband Users Affected 0.09%
Started 2 Feb 10:09:12 by AAISP automated checking
Closed 16 Apr 15:59:33

13 Apr 15:01:38
13 Apr 14:51:55
There was an issue with two of our routers - a few lines dropped, and are reconnecting. Routing was affected for a minute or two. We're investigating.
Resolution Service has recovered as expected. We'll see if we can find the underlying cause. Sorry for any inconvenience.
Started 13 Apr 14:46:46
Closed 13 Apr 15:01:38
Previously expected 13 Apr 14:50:00

7 May 08:33:30
2 Apr 15:48:08
We expect to do some router upgrades, including normal rolling LNS upgrades over the next week as a new release of the FireBrick is expected to be released shortly. This should have little or no disruption, as usual.
9 Apr 12:52:05
This was a bit delayed and should start tonight, and be ongoing in to the weekend.
25 Apr 09:08:02
Further updates this weekend (25/26)
Started 3 Apr
Closed 7 May 08:33:30
Previously expected 1 May

2 Apr 16:02:22
17 Mar 12:38:27
We are seeing higher than normal evening time latency on the Wrexham exchange, it is not every night but it does suggest BT are running another congested link. This has been reported to them and we will update thia as and when they get back to us.
17 Mar 12:41:51
Here is an example graph:
20 Mar 14:36:18
It has looked better the last two eveings but it's still being investigated as the BT links were probably less busy.
Broadband Users Affected 0.01%
Started 15 Mar 12:36:07 by AAISP Staff
Closed 2 Apr 16:02:22

2 Apr 11:57:32
1 Apr 10:00:06
Some customers connected through Gloucestershire are affected by an ongoing TalkTalk major service outage. Details below: Summary

Network monitoring initially identified total loss of service to all customers connected to 2 exchanges in the Gloucester area. Our NOC engineers re-routed impacted traffic whilst virgin media engineers carried out preliminary investigations. Virgin media restoration work subsequently resulted in several major circuits in the Gloucester area to fail.

This has resulted in a variety of issues for multiple customers connected to multiple exchanges. Our NOC engineers have completed re-routing procedures to restore service for some customers with other customers continuing to experience total loss of service due to capacity limitations. Impact: Tigworth, Witcombe and Painswick exchanges

Hardwicke and Barnwood exchanges – experiencing congestion related issues.

Cheltenham and Churchdown – experiencing congestion related issues.

experiencing congestion related issues Stroud, Stonehouse, Whitecroft, Blakeney, Lydney, Bishops Cleeve, Winchcombe, Tewkesbury, Bredon exchanges.

1 Apr 10:31:30
TT have advised that splicing of the affected fibre is still ongoing. There is no further progressive updates at this time. Further updates will be sent out shortly.
2 Apr 11:57:26
Root cause analysis identified a major Virgin Media fibre break due to third party contractor damage as being the cause of this incident. Service was fully restored when Virgin Media Fibre engineers spliced new fibre. Following this we received confirmation that service had returned as BAU. TalkTalk customers would have been automatically rerouted and would have experienced only a momentary Loss of Service. An observation period has been carried out verifying network stability and as no further issues have been reported this incident will be closed with further investigations into the cause being tracked via the Problem Management process
Closed 2 Apr 11:57:32

27 Mar 09:00:00
25 Mar 21:48:13
Since the 24th March we have been seeing congestion on TalkTalk lines on the Shepherds Bush exchange. This has been reported to TalkTalk. Example graph:
26 Mar 10:51:48
TalkTalk say they have fixed this. We'll be checking overnight to be sure smiley
26 Mar 22:27:33
Lines are looking good.

We had this feedback from TalkTalk regarding this congestion issue:

Shepherds Bush has three GigE backhauls to two different BNG's - there was some a software process failure and restart on one of these devices on Tuesday morning which had two of the three backhauls homed to it. As a result all customers redialled to the one 'working' BNG in the exchange - normally when this happens we will calculate whether or not the backhaul can handle that number of customers and if not manually intervene, in this case however a secondary knock on issue meant that our DHCP based customers (FTTC subs) were sent through the same backhaul and the calculation was inaccurate.

If the PPP session was restarted they would have reconnected on their normal BNG and everything should be OK - we've just made this change manually moving subscribers over - still have a couple of lines on the backup BNG so will monitor if there are any issues and take any necessary actions to resolve.

Started 24 Mar 17:00:00
Closed 27 Mar 09:00:00

17 Mar 11:18:55
20 Jan 12:53:37
We are seeing low level packet loss on some BT circuits connected to the EUSTON exchange, this has been raised with BT and as soon as we have an update we will post an update here.
20 Jan 12:57:32
Here is an example graph:
22 Jan 09:02:48
We are due an update on this one later this PM
23 Jan 09:36:21
BT are chasing this and we are due an update at around 1:30PM.
26 Jan 09:41:39
Work was done over night on the BT side to move load onto other parts of the network, we will check this again this evening and report back.
27 Jan 10:33:05
We are still seeing lines with evening packet loss but BT don't appear to understand this and after spending the morning arguing with them they have agreed to investigate further. Update to follow.
28 Jan 09:35:28
Update from BT due this PM
29 Jan 10:33:57
Bt are again working on this but no further updates will be given until tomorrow morning
3 Feb 16:19:06
This one has also been escalated further with BT
4 Feb 10:18:11
BT have identified a fault within their network and we have been advised that an update will be given after lunch today
11 Feb 14:16:56
Yet another rocket on it's way to BT
24 Feb 12:59:20
escalated further with BT, update due by the end of the day.
2 Mar 09:59:19
STill waiting for BT to raise an emergency PEW, the PEW (planned engineering work) will sort the last few lines where we are seeing peak time packet loss)
12 Mar 12:03:57
I need to check this tonight as bT think it is fixed, I will post an update back tomorrow
Broadband Users Affected 0.07%
Started 10 Jan 12:51:26 by AAISP automated checking
Closed 17 Mar 11:18:55
Previously expected 21 Jan 16:51:26

13 Mar 22:57:53
13 Mar 22:44:00
TalkTalk lines lost connection at about 10:30, and are reconnecting at the moment.
13 Mar 22:47:49
About 80% of lines have now reconnected.
13 Mar 22:58:06
Most lines are back.
Resolution Confirmed as a fault within the TalkTalk network that affected us and other ISPs.
Started 13 Mar 22:30:00
Closed 13 Mar 22:57:53