Over the retiring twelvemonth , we have seena manna from heaven in artificial intelligent ( AII ) engineering . From the good(choosing viable IVF embryos)to the bad(designing inspirational posters)to the strange(coming up with quaint – sometimes rude – British place name ) , its success has been a bit of a mixed bag .

The latest news in the world of AI comes from London , UK , where the city ’s Metropolitan Police have announced plans to practice artificial intelligence to scan electronic devices for images of child abuse . They say the engineering science should be quick for use within"two to three years " .

As of the right way now , however , there is just one small singultus   – the software mistakes desert landscape for nude photographs . presumptively , it ’s because of the colouring .

The Met already use a less elegant form of image recognition software . The job is - as Mark Stokes explain in an consultation withThe Telegraph – while the software program is capable to   pick up   certain forms of criminal activity ( shooter , drugs , and money ) , it has a much unvoiced clock time name pornographic ikon and videos . Stokes is the Metropolitan Police ’s   head of digital and electronic forensics .

This means it ’s up to constabulary officer themselves to go through the indecent images and mark them for unlike sentencing - a grueling and psychologically stressful exercise , especially given the plate of the task . According toThe Telegraph , the Met had to research through 53,000 devices for imply images in 2016 . In February , a direct   police policeman send for the levels of recorded tiddler sex   offensive activity in the country"staggering " .

" you may suppose that doing that for twelvemonth - on - year is very disturbing , " Stokes told the Telegraph .

Fortunately , technological improvements could intend the task is excrete on to inanimate object , who wo n’t be psychologically affected by the job . The British constabulary force out is presently working with Silicon Valley providers ( such as Google and Amazon ) to create AI technology advanced enough to identify abusive imagery but there are still a few glitches to smoothen over .

" Sometimes it come up with a desert and it thinks it ’s an indecent picture or pornography , ” Stokes explained at the ( ISC)2 Secure Summit in London .

Which is problematic because desert landscape   are a popular screensaver and wallpaper choice .

" For some reason , lots of people have concealment - savers of comeuppance and it picks it up consider it is skin colour . "

[ H / T : The telegraphy ]