This affects SIP2SIM customers who would previously have seen short CLI on calls to their SIM - eg where an incoming call to the SIM would have shown the short extension number of a caller on the same PABX.
Our SIP2SIM service is intended to allow a SIM in any phone or any 'dumb' phone to work as a SIP extension, either a SIP number provided by A&A or an extension on a customer's own SIP PBX. This meant that, for example, for an internal call, you would see the call from that internal extension number. This was not perfect, e.g. a call from extn 401 might show as +401, but it worked.
A change in Ofcom policy has meant that we now have to send a 'valid' looking CLI SIM. This is fine for normal calls from 'outside' your PABX as they will have the proper CLI of the caller, but customers running their own PBX may well be using short extension numbers, and 'internal' calls on the customer PBX between extensions would be sending the extension number as the CLI. eg a three digit extension number such as 402 or 403 etc.
In situations where an invalid CLI is used - eg an 'internal' call to the SIP2SIM extension, we have to now change the CLI to look like a 'valid' CLI.
Our workaround is to change the CLI to be: +11, then a number of zeros, then the original CLI requested, so +11 and 9 further digits.
eg, if you have two 'internal' extension numbers, 402 and 403. 402 is a normal SIP phone, and 403 is registered from a SIP2SIM. A call from 402 to 403 will have the CLI changed to: +11110000402
This does mean a call back to such a number would hit your PABX with this otherwise invalid number. We suggest you make your PABX map these to internal extensions if possible.
This does somewhat break the philosophy of our SIP2SIM service as a SIM being a normal SIP extension. We are not happy about this. This change is mandated by Ofcom and we have no choice.
The graph shows the last few hours of logins and logouts of ADSL, VDSL, SIMs and L2TP circuits.
The current time is on the left. Green is login, red is logout.
If there are spikes, then this shows a large number of logouts, which may indicate an outage or planned work happening.
You can click on a spike to search for incidents or maintenance that were open around that time.
This will search through the various outages and maintenance reports we have received from our suppliers which may be affecting you.
You can also see a list of outages and maintenance work we receive from our suppliers that may affect your specific broadband service on the Control Pages.
The search box above should find posts affecting your service however, you can use a Keyword search.
This is the status page of Andrews & Arnold Ltd.
Our status page shows outages (problems) and maintenance (planned work) that happen on our own network and systems and also that of our suppliers networks and systems. We try and ensure this site is updated as soon as possible with incidents as they happen. Live discussion of issues is usually available on IRC.
The last update was 11 Sep 09:10:16
Every night we analyse the packet loss and latency from us to our carriers and also from us to each of our customers.
This table show un-errored seconds on our connections in to our carriers for the past 7 days. Our aim is for all of these to be at 100%.
We have links to back-haul carriers BT and TT.
Our target is that we operate a uncongested links to the carriers. This means that if the general trend is that the link will be getting full we order more capacity from carriers. The capacity we order is expensive, takes time to change and has minimum terms. This means it can be difficult to perfectly manage the traffic. We allow some head room but not more than necessary. It is technically impossible to guarantee that the link is always uncongested (not without having a pipe the size of all internet links in the world added together) but by careful monitoring and allowing enough headroom we can aim for that target.
This table shows an analysis of how well we meet our target.
The unerrored seconds report is the simplest statistic to understand. If we drop even a single packet in a second then this counts as an errored second. Given that we can be handling hundreds of thousands of packets every second, this can be a very sensitive measure of congestion. Bear in mind we will drop the larger packets first which will normally be TCP which re-sends the dropped plackets. This helps ensure VoIP and interactive uses of the internet are unaffected.
Daytime unerrored seconds are based only on 9am to 6pm Monday to Friday
We also consider 100 second samples. If we have reduced the throughput of non-premium lines at all (as a result of dropped packets for several seconds in a row) then the whole 100 seconds is considered to have affected non-premium customers. Because we are looking at 100 seconds at a time it is possible for very short busts of traffic to cause this to show a worse figure than the unerrored seconds. Even if traffic was reduced by 0.5% for part of the 100 second period the whole 100 seconds counts as errored.
Considering the 100 seconds samples, there is a limit to how much we will reduce the non-premium customers. If that limit is reached then we consider the whole 100 second sample is having an impact on premium lines. This is a sign of much more severe bursts of congestion for several minutes. Bear in mind, a single 100 second sample in a day is more than 0.1% of the day, and even if this does happen VoIP and interactive services should be unaffected.Read more.
|Link||Week days 9-6||24 hour||Non-premium||Premium|
This report can pinpoint congestion in exchanges, BRAS's or particular areas of the country.
Every line is sent an LCP echo every second and loss and latency measured. This is summarised per 100 second sample and archived for every line for each whole day. The previous day is then analysed each morning and the above table is updated.
Each line is considered to pass, fail or be inconclusive. Inconclusive includes lines where samples have too much upload or download which could themselves cause higher loss or latency to be observed. Lines with less than 18 hours on-line are excluded totally.
Latency is considered against a reference base-line latency for the circuit, and this has to be present for several hours to be considered. As such, lines without clean latency monitoring (some makes of router) are excluded.
For an area to show in the above table there must be no lines that pass at all, and at least 50% of the lines must fail (the rest being inconclusive). Where we have very few lines we require 80% to fail (and none to pass) to show on this list. The area must also have enough lines, so some smaller exchanges will not be detected.
The size stated is what percentage of our total lines are in the affected area.
The block graph under each report for yesterday is based on an average of valid samples for each line for each hour, and then taking the lowest value of all lines in the affected area, so the loss and latency shown is something that is affecting every line in an area.
There are no currently identified congestion spots
We are not able to automatically identify all congestion spots so this does not mean there are not problem areas, do contact Support if you think you are suffering from congestion.