Tag Archives: DWP

Data illusions

Yesterday was an interesting day for a few reasons; one of the primary reasons was an opinion piece in the Guardian by Jay Watts (@Shrink_at_Large). Like many article I considered to be in opposition, yet when I reread it, this piece has all kinds of hidden gems and I had to ponder a few items for an hour or so. I love that! Any piece, article or opinion that makes me rethink my position is a piece well worth reading. So this piece called ‘Supermarkets spy on them now‘ (at https://www.theguardian.com/commentisfree/2018/may/31/benefits-claimants-fear-supermarkets-spy-poor-disabled) has several sides that require us to think and rethink issues. As we see a quote like “some are happy to brush this off as no big deal” we identify with too many parts; to me and to many it is just that, no big deal, but behind the issues are secondary issues that are ignored by the masses (en mass as we might giggle), yet the truth is far from nice.

So what do we see in the first as primary and what is behind it as secondary? In the first we see the premise “if a patient with a diagnosis of paranoid schizophrenia told you that they were being watched by the Department for Work and Pensions (DWP), most mental health practitioners would presume this to be a sign of illness. This is not the case today.” It is not whether this is true or not, it is not a case of watching, being a watcher or even watching the watcher. It is what happens behind it all. So, when we recollect that dead dropped donkey called Cambridge Analytics, which was all based on interacting and engaging on fear. Consider what IBM and Google are able to do now through machine learning. This we see in an addition to a book from O’Reilly called ‘The Evolution of Analytics‘ by Patrick Hall, Wen Phan, and Katie Whitson. Here we see the direct impact of programs like SAS (Statistical Analysis System) in the application of machine learning, we see this on page 3 of Machine Learning in the Analytic Landscape (not a page 3 of the Sun by the way). Here we see for the government “Pattern recognition in images and videos enhance security and threat detection while the examination of transactions can spot healthcare fraud“, you might think it is no big deal. Yet you are forgetting that it is more than the so called implied ‘healthcare fraud‘. It is the abused setting of fraud in general and the eagerly awaited setting for ‘miscommunication’ whilst the people en mass are now set in a wrongly categorised world, a world where assumption takes control and scores of people are now pushed into the defence of their actions, an optional change towards ‘guilty until proven innocent’ whilst those making assumptions are clueless on many occasions, now are in an additional setting where they believe that they know exactly what they are doing. We have seen these kinds of bungles that impacted thousands of people in the UK and Australia. It seems that Canada has a better system where every letter with the content: ‘I am sorry to inform you, but it seems that your system made an error‘ tends to overthrow such assumptions (Yay for Canada today). So when we are confronted with: “The level of scrutiny all benefits claimants feel under is so brutal that it is no surprise that supermarket giant Sainsbury’s has a policy to share CCTV “where we are asked to do so by a public or regulatory authority such as the police or the Department for Work and Pensions”“, it is not merely the policy of Sainsbury, it is what places like the Department for Work and Pensions are going to do with machine learning and their version of classifications, whilst the foundation of true fraud is often not clear to them, so you want to set a system without clarity and hope that the machine will constitute learning through machine learning? It can never work, that evidence is seen as the initial classification of any person in a fluidic setting is altering on the best of conditions. Such systems are not able to deal with the chaotic life of any person not in a clear lifestyle cycle and people on pensions (trying to merely get by) as well as those who are physically or mentally unhealthy. These are merely three categories where all kind of cycles of chaos tend to intervene with their daily life. Those are now shown to be optionally targeted with not just a flawed system, but with a system where the transient workforce using those methods are unclear on what needs to be done as the need changes with every political administration. A system under such levels of basic change is too dangerous to get linked to any kind of machine learning. I believe that Jay Watts is not misinforming us; I feel that even the writer here has not yet touched on many unspoken dangers. There is no fault here by the one who gave us the opinion piece, I personally believe that the quote “they become imprisoned in their homes or in a mental state wherein they feel they are constantly being accused of being fraudulent or worthless” is incomplete, yet the setting I refer to is mentioned at the very end. You see, I believe that such systems will push suicide rates to an all-time high. I do not agree with “be too kind a phrase to describe what the Tories have done and are doing to claimants. It is worse than that: it is the post-apocalyptic bleakness of poverty combined with the persecution and terror of constantly feeling watched and accused“. I believe it to be wrong because this is a flaw on both sides of the political aisle. Their state of inaction for decades forced the issue out and as the NHS is out of money and is not getting any money the current administration is trying to find cash in any way that they can, because the coffers are empty, which now gets us to a BBC article from last year.

At http://www.bbc.com/news/election-2017-39980793, we saw “A survey in 2013 by Ipsos Mori suggested people believed that £24 out of every £100 spent on benefits was fraudulently claimed. What do you think – too high, too low?
Want to know the real answer? It is £1.10 for every £100
“. That is the dangerous political setting as we should see it; the assumption and believe that 24% is set to fraud when it is more realistic that 1% might be the actual figure. Let’s not be coy about it, because out of £172.3bn a 1% amount still remains a serious amount of cash, yet when you set it against the percentage of the UK population the amount becomes a mere £25 per person, it merely takes one prescription to get to that amount, one missed on the government side and one wrongly entered on the patients side and we are there. Yet in all that, how many prescriptions did you the reader require in the last year alone? When we get to that nitty gritty level we are confronted with the task where machine learning will not offer anything but additional resources to double check every claimant and offense. Now, we should all agree that machine learning and analyses will help in many ways, yet when it comes to ‘Claimants often feel unable to go out, attempt voluntary work or enjoy time with family for fear this will be used against them‘ we are confronted with a new level of data and when we merely look at the fear of voluntary work or being with family we need to consider what we have become. So in all this we see a rightful investment into a system that in the long run will help automate all kinds of things and help us to see where governments failed their social systems, we see a system that costs hundreds of millions, to look into an optional 1% loss, which at 10% of the losses might make perfect sense. Yet these systems are flawed from the very moment they are implemented because the setting is not rational, not realistic and in the end will bring more costs than any have considered from day one. So in the setting of finding ways to justify a 2015 ‘The Tories’ £12bn of welfare cuts could come back to haunt them‘, will not merely fail, it will add a £1 billion in costs of hardware, software and resources, whilst not getting the £12 billion in workable cutbacks, where exactly was the logic in that?

So when we are looking at the George Orwell edition of edition of ‘Twenty Eighteen‘, we all laugh and think it is no great deal, but the danger is actually two fold. The first I used and taught to students which gets us the loss of choice.

The setting is that a supermarket needs to satisfy the need of the customers and the survey they have they will keep items in a category (lollies for example) that are rated ‘fantastic value for money‘ and ‘great value for money‘, or the top 25th percentile of the products, whatever is the largest. So in the setting with 5,000 responses, the issue was that the 25th percentile now also included ‘decent value for money‘. So we get a setting where an additional 35 articles were kept in stock for the lollies category. This was the setting where I showed the value of what is known as User Missing Values. There were 423 people who had no opinion on lollies, who for whatever reason never bought those articles, This led to removing them from consideration, a choice merely based on actual responses; now the same situation gave us the 4,577 people gave us that the top 25th percentile only had ‘fantastic value for money‘ and ‘great value for money‘ and within that setting 35 articles were removed from that supermarket. Here we see the danger! What about those people who really loved one of those 35 articles, yet were not interviewed? The average supermarket does not have 5,000 visitors, it has depending on the location up to a thousand a day, more important, when we add a few elements and it is no longer about supermarkets, but government institutions and in addition it is not about lollies but Fraud classification? When we are set in a category of ‘Most likely to commit Fraud‘ and ‘Very likely to commit Fraud‘, whilst those people with a job and bankers are not included into the equation? So we get a diminished setting of Fraud from the very beginning.

Hold Stop!

What did I just say? Well, there is method to my madness. Two sources, the first called Slashdot.org (no idea who they were), gave us a reference to a 2009 book called ‘Insidious: How Trusted Employees Steal Millions and Why It’s So Hard for Banks to Stop Them‘ by B. C. Krishna and Shirley Inscoe (ISBN-13: 978-0982527207). Here we see “The financial crisis appears to be exacerbating fraud by bank employees: a new survey found that 72 percent of financial institutions say that in the last 12 months they have experienced a case of data theft by one of their workers“. Now, it is important to realise that I have no idea how reliable these numbers are, yet the book was published, so there will be a political player using this at some stage. This already tumbles to academic reliability of Fraud in general, now for an actual reliable source we see KPMG, who gave us last year “KPMG survey reveals surge in fraud in Australia“, with “For the period April 2016 to September 2016, the total value of frauds rose by 16 percent to a total of $442m, from $381m in the previous six month period” we see number, yet it is based on a survey and how reliable were those giving their view? How much was assumption, unrecognised numbers and based on ‘forecasted increases‘ that were not met? That issue was clearly brought to light by the Sydney Morning Herald in 2011 (at https://www.smh.com.au/technology/piracy-are-we-being-conned-20110322-1c4cs.html), where we see: “the Australian Content Industry Group (ACIG), released new statistics to The Age, which claimed piracy was costing Australian content industries $900 million a year and 8000 jobs“, yet the issue is not merely the numbers given, the larger issue is “the report, which is just 12 pages long, is fundamentally flawed. It takes a model provided by an earlier European piracy study (which itself has been thoroughly debunked) and attempts to shoe-horn in extrapolated Australian figures that are at best highly questionable and at worst just made up“, so the claim “4.7 million Australian internet users engaged in illegal downloading and this was set to increase to 8 million by 2016. By that time, the claimed losses to piracy would jump to $5.2 billion a year and 40,000 jobs” was a joke to say the least. There we see the issue of Fraud in another light, based on a different setting, the same model was used, and that is whilst I am more and more convinced that the European model was likely to be flawed as well (a small reference to the Dutch Buma/Stemra setting of 2007-2010). So not only are the models wrong, the entire exercise gives us something that was never going to be reliable in any way shape or form (personal speculation), so in this we now have the entire Machine learning, the political setting of Fraud as well as the speculated numbers involved, and what is ‘disregarded’ as Fraud. We will end up with a scenario where we get 70% false positives (a pure rough assumption on my side) in a collective where checking those numbers will never be realistic, and the moment the parameters are ‘leaked’ the actual fraudulent people will change their settings making detection of Fraud less and less likely.

How will this fix anything other than the revenue need of those selling machine learning? So when we look back at the chapter of Modern Applications of Machine Learning we see “Deploying machine learning models in real-time opens up opportunities to tackle safety issues, security threats, and financial risk immediately. Making these decisions usually involves embedding trained machine learning models into a streaming engine“, that is actually true, yet when we also consider “review some of the key organizational, data, infrastructure, modelling, and operational and production challenges that organizations must address to successfully incorporate machine learning into their analytic strategy“, the element of data and data quality is overlooked on several levels, making the entire setting, especially in light of the piece by Jay Watts a very dangerous one. So the full title, which is intentionally did not use in the beginning ‘No wonder people on benefits live in fear. Supermarkets spy on them now‘, is set wholly on the known and almost guaranteed premise that data quality and knowing that the players in this field are slightly too happy to generalise and trivialise the issue of data quality. The moment that comes to light and the implementers are held accountable for data quality is when all those now hyping machine learning, will change their tune instantly and give us all kinds of ‘party line‘ issues that they are not responsible for. Issues that I personally expect they did not really highlight when they were all about selling that system.

Until data cleaning and data vetting gets a much higher position in the analyses ladder, we are confronted with aggregated, weighted and ‘expected likelihood‘ generalisations and those who are ‘flagged’ via such systems will live in constant fear that their shallow way of life stops because a too high paid analyst stuffed up a weighting factor, condemning a few thousand people set to be tagged for all kind of reasons, not merely because they could be optionally part of a 1% that the government is trying to clamp down on, or was that 24%? We can believe the BBC, but can we believe their sources?

And if there is even a partial doubt on the BBC data, how unreliable are the aggregated government numbers?

Did I oversimplify the issue a little?

 

 

Leave a comment

Filed under Finance, IT, Media, Politics, Science

They don’t know what they do!

The article started funny enough. The headline ‘Leaked universal credit memo shows jobcentre staff struggling with rollout’ gave me a clear indication that this is another one of these, let’s get into a world we do not understand (at http://www.theguardian.com/uk-news/2014/oct/27/universal-credit-leaked-memo-scheme-rollout).

I admit that my words here are presumptuous, but I have seen this before, to be honest many of us have seen this before. There was the NHS with 14 billion plus wasted and there were a few other projects, all gone down the drain. So, why can’t some people get their act together?

The first quote is likely the most offensive one, especially in my eyes: “The DWP had promised to have 1 million people on the scheme by April 2014 but, dogged by delays and tens of millions of pounds of IT write-downs and write-offs, the original timetable has been scrapped. Just 15,000 people are on the system“, you think what is wrong with this picture. Consider a $389 notebook, not a great piece of equipment, but I can install a variety of SQL products and have these filled with a database containing the population data of Poland in about an hour, so why do we see a system with only 15,000 records? (intentional trivialisation was used here)

When we get to the timeline (which by the way was not chronological), we see several issues. Let us take a look at them.

28th April 2013 – Trial begins for Universal credit (UC), which is covered in the Welfare Reform Act 2012 (at http://www.theguardian.com/politics/2013/mar/31/liberal-conservative-coalition-conservatives)

Universal credit introduced.
The new in- and out-of-work credit, which integrates six of the main out-of-work benefits, will start to be implemented this April in one jobcentre in Ashton-under-Lyne, Greater Manchester. The aim is to increase incentives to work for the unemployed and to encourage longer hours for those working part-time. It had been intended that four jobcentres would start the trial in April, but this has been delayed until July, and a national programme will start in September for new claimants. They will test the new sanctions regime and a new fortnightly job search trial, which aims to ensure all jobseeker’s allowance and unemployment claimants are automatically signed onto Job Match, an internet-based job-search mechanism. Suspicion remains that the software is not ready.

The issues are as follows:

  1. Will start to be implemented this April‘, this means that the system had been prototyped, this means that the software has been tested and that the interface has been tested by users, so that a nearly clean version goes online.
  2. The information ‘Suspicion remains that the software is not ready‘, should have been a very clear indication that the brakes had to be applied and at this point, investigations on the entire track should have commenced.

24th May 2013 The Major Projects Authority review expresses serious concerns about the department having no detailed “blueprint” and transition plan for UC (at http://www.theguardian.com/society/2013/may/24/universal-credit-danger-failing-whitehall-review)

Universal credit in danger of failing, official Whitehall review says
The first official government admission that Iain Duncan Smith’s flagship plans to remake the welfare state has hit trouble emerged on Friday night when the Cabinet Office’s review of all major Whitehall projects branded the universal credit programme as having fallen into “amber-red” status, a category designating a project in danger of failing.

You think? How about, the issues shown after a month when there were already doubts we see an utter lack of commitment, there is no other way to describe it. When I see the quote “Francis Maude, the Cabinet Office minister, hailed the publication: ‘Major projects need scrutiny and support if we are to succeed in the global race’“, which in my book comes across as ‘only silent scrutiny is allowed. This project is too big‘, which in my eyes is nothing less than a joke, one the taxpayer is paying for by the way. I must also clarify that this is how I initially read it, not how Francis Maude stated it, he seems to want accountability, so do I, it is just too convenient that many involved are not named at all.

In addition we see “An MPA rating of amber-red will anger the DWP, which has insisted that universal credit is on time and on budget” furthermore we see “Data has been exempted from only 21 projects in the review by the Major Projects Authority (MPA), where disclosure would damage commercial interests or national security“.

So now we get the following:

  1. Who at the DWP had made that statement? We want to see his name and his dismissal; I say again dismissal, not his resignation.
  2. Was the same person making the claims in regards to October 2013? This means that we were at that point faced with two delays on a pretty expensive endeavour. More important, until now, there has been a slacking handle on this project, which is likely to be only one of many.

Now we look at two events:

5th September 2013 A National Audit Office report reveals ministers have written off £34m on failed IT programmes and the launch may be delayed beyond 2017 (at http://www.theguardian.com/politics/2013/sep/05/david-cameron-24bn-universal-credit-problems), where we see ‘David Cameron’s £2.4bn universal credit project riddled with problems‘, so the entire UC is more than just a few pennies and we are not seeing any accountability, no criminal charges and no product. We can look at the quote, which is “The National Audit Office said universal credit, the £2.4bn project meant to consolidate six welfare payments into one, has been beset by ‘weak management, ineffective control and poor governance’“, I am about to call it something else entirely.

31st October 2013 The Guardian reveals ministers have been presented with a radical plan to restart UC and write off £119m of work over the past three years (at http://www.theguardian.com/politics/2013/oct/31/universal)

Now we see the following additional quotes “Ministers attempting to put the troubled universal credit welfare reform programme back on track have been presented with a radical plan to restart the scheme and write off £119m of work over the past three years” and “The risk assessment warns that the plan to start again, the ‘design and build’ web-based scheme, is ‘unproven … at this scale’“. It says the plan to fix three years of work on universal credit is still “not achievable within the preferred timescales“, describing it as unrealistic”

These two give us the following:

  1. If we revisit “In March 2013 Duncan Smith told parliament that universal credit ‘is proceeding exactly in accordance with plans’“, then why on earth is Duncan Smith in any government building? If we look at statements from Margaret Hodge and the NAO, there is a clear indication that extreme sanitisation is needed at the DWP, the fact that this multi-billion pound fiasco is still around at that time should give cause to many serious questions.

Just to make sure the reader understands the gravity of this situation, the bungling and wasting of resources at that point could have given nearly every current university student a FREE University degree, which is saying a lot, in addition, those studying IT, might have completed the project for the price of their education, which is saying a lot!

  1. Writing off 119 million of work delivered. A failure is not work delivered, who was minding the stores, the contracts as well as the targets that had to be met? The fact that the amount in the database at present (15,000 people) could have been achieved with a $99 program called Microsoft Access, so can we have the 118,999,901 back please?

When we revisit the September quote “The DWP said the department would continue with the planned reform and was committed to delivering it on time by 2017 and within budget“, we can clearly see that either the DWP has no clue what it was doing, or we have another echelon of people and their ‘goals‘ messing things up.

Are my assumptions valid? Well, so far I did not waste billions, so I am inclined to say yes!

By the way, who did the original costing, who presented the plan and what remains of the initial plan? Because a blowout of these proportions should be regarded as clear evidence that the thought might have been nice, but none of the deciding parties had any clue on what was being decided on (my evidence here are the squandered billions as we see them melting away).

You see, in the old days, in my life, designing a database system was relatively simple. It took 5 weeks and a few iterations of tweaking to get the customer this container system. It worked like a charm! That is what is needed here. People have been overcomplicating things by massive portions.

  1. Web based solution.
    Really? With all the intrusions, phishing and other forms of malignant issues, you are going to a web based option? Let’s be clear, this system is all about letters and numbers, so an ASCII based system, which in the old days it was called a DOS program. In this situation a UNIX solution should be sought, but the overall idea is clear. In addition, UNIX is much safer, better protected and scripting allows for evolution when needed. I knew a guy once, who created a scripted solution for product distribution for a global Fortune 500 company, it was one of the few innovative software solutions that actually worked and worked when most systems had to be upgraded, it worked on a Pentium 1 with 90 MHz, a system we now buy for $49 (if even for that much), It conversed with several dozens of locations.

Now, today, when we look at the UC, something bigger is needed, but the systems of yesterday are already 2000 times stronger than the initial system it was designed on, so we can clearly see that the spending of a few billion require a deeper digging, as well as a serious interview by the members of the House of Commons towards the involved members of the DWP.

  1. more web-based system
    The risk assessment, dated 11th October, says the plan for a faster, more web-based system would involve writing off £119m of previous work, and cost the DWP £96m to develop. However, it warns ministers that they will have no idea if the web-based system will work until the summer of 2014 ‘when it is live for 100 claimants’

And the laughter just does not stop here, ‘more’ web based system? The people here did not learn the first time? If you want speed, consider simple ASCII, with perhaps local formatted XML. You see, you get loads of characters across in mere milliseconds (36 characters including 10 numbers tends to be fast), and let us not forget, this is all set towards 6 systems, so you need speed. So only this summer was there any chance of knowing anything, so can we wonder again where the money went, because someone is getting pretty rich here and it is not me (alas).

In these two issues we see a reiterated failure, which gives a clear signal that the original design, which would have been BEFORE money was spend, should not have passed any hurdles as I see it.

When I think ANY project I see the following

  1. request
  2. design
  3. prototype
  4. finalise
  5. test
  6. implement

Now, I will admit that a large project needs a lot more, but these 6 steps for the initial trial should have been done in 90 days for 7 tests. One test of each system and the 7th to see one person collected on all 6 systems. Now we have a master that gets us trials where this simple program could be used to star testing everywhere and see if data comes across, yes, this is nowhere near finished, but in the foundation we see what happens if the data of 150,000 people gets requested, so now we know that data can be obtained and we see a timeline of speed and more important bandwidth, because that will be the killer. If we revisit the original time line where the plan was offered in October 2010, which means that this test could have been done before Christmas, so how was time and money wasted, because as we see the Multi Billion pound bill that would be the direct question evolving from this.

The complications
Yes, I am not ignoring this. A system with this much data access will need all levels of security and encryption, there is no denying this, yet using a ‘web-based’ approach seems to me that we might as well give a copy of all this data to the cyber criminals. There are always suite options of security, and yes that needs work, yet some local test could have been made, in addition, a system this vast will need all kind of implementation servers and trained support staff, steps that were not even anywhere near implementing, were they costed for?

When we see the timeline and the involvement from ‘interested’ parties, I cannot stop but wonder what could have been if the right people had sat down, because those involved screwed the pooch big time and the taxpayer can see the billions they have to cough up for a system that never worked.

We will end with three quotes all from the October 27th 2014 article.

  1. leaked Whitehall documents warned of a failing IT system, more than £1m in wasted expenditure, and how only 25,000 claimants would likely to be served by the system by the general election next year.
  2. The government has written off or written down £130m on the project, which is designed to revolutionise the culture around claiming benefits. It now expects 100,000 people to be on the system by May 2015 and for 100 centres to be involved in its delivery by the end of this year.
  3. When fully rolled out, UC will make 3 million families better off by £177 a month and lift up to 300,000 children out of poverty.”

From the three points we get the following, if the system is turning nuts and bolts at present when there are between 25,000 and 100,000, what complications will we see when the other 2.95 million are added, if we see the issues with less than 4% populated, what happens when the other 96% is added?

When we see the quote in regards to a couple not getting paid, whilst in addition changing their details took three months, we can conclude form the quote “The DWP said the couple’s claim had been delayed because the pair had failed to complete the correct forms. Responding to Dispatches’ findings, a spokesman told the Guardian: ‘Universal credit’s IT system is robust and effective, and we have trained 26,300 work coaches who are successfully providing new support to claimants to help them better prepare for work’“, well if there are 26,300 work coaches and there are currently 25,000 in the system, why did it take three months to correct this? In addition, how come the wrong forms were filled in, what was the cause of that? Should the system not have reported (almost immediately) that the forms did not constitute their current social status/predicament?

This is more than a simple failing; this system seems to lack basic foundations, especially with three months delays.

The sad part is that this is not the first issue we see, when we consider the NHS debacle which I discussed in ‘the second exploitation‘ on August 10th, how the NHS options resulted in a wasted 15 billion, whilst no one seems to take a deeper look at how such large amounts get wasted. Now with the UC we see a similar development, it would be so nice for someone in Whitehall to recognise the need for actual change so that squandering might be minimised be a lot more then it currently is.

Leave a comment

Filed under Finance, IT, Law, Media, Politics, Science