Hi Friends,

Even as I launch this today ( my 80th Birthday ), I realize that there is yet so much to say and do.

There is just no time to look back, no time to wonder,"Will anyone read these pages?"

With regards,
Hemen Parekh
27 June 2013

Friday, 4 August 2017

Racing towards ARIHANT ?





Today { year 2017 } :


Mumbai Mirror ( 03 Aug ) carries following news report :



“ Robots tackle extremist videos better : Google “


Google has said that its artificial intelligence (AI)-driven robots are more accurate than humans in identifying and blocking the extremist videos online.


According to a report in The Telegraph on Tuesday, Google claimed that its robots flag three in four offensive videos from 
YouTube even before they are reported by users.


"With over 400 hours of content uploaded to YouTube every minute, finding and taking action on violent extremist content poses a significant challenge," the report quoted Google as saying.



"But over the past month, our initial use of machine learning has more than doubled both the number of videos we've removed for violent extremism, as well as the rate at which we've taken this kind of content down," the company added.


The AI-driven bots at Google which are meant to identify the offensive content have identified more than 75 per cent of videos removed from YouTube in July.



The company sees this ability of bots as "dramatic" and better than humans.


Facebook is also using AI to remove terrorist propaganda and related content, using image recognition software and video technology.


Social media companies and other 
Internet companies have been under scanner for doing little to curb the spread of violence and terror-related content on their platforms.




Flashback { year 2013 } :



Eric Schmidt and Jared Cohen ( both from Google ) wrote in :  THE  NEW  DIGITAL  AGE  ( page : 181 ) :


We can already see early steps in this direction, as some companies make clear statements in policy and procedure


At YouTube, there is the challenge of content volume . With more than four billion videos viewed daily ( and sixty hours of video uploaded each minute ), it is impossible for the company to screen that content for what is considered inappropriate material, like advocating terrorism


Instead YouTube relies on a process in which users flag content they consider inappropriate ; the video in question goes to a YouTube team for review, and it is taken down if it violates company policies


Eventually, we will see the emergence of industry-wide standards


All digital platforms will forge a common policy with respect to dangerous extremist videos online, just as they have coalesced in establishing policies governing child pornography


There is a fine line between censorship and security, and we must create safeguards accordingly



THE INDUSTRY WILL WORK AS A WHOLE TO DEVELOP SOFTWARE THAT MORE EFFECTIVELY IDENTIFIES VIDEOS WITH TERRORIST CONTENT



Some in the industry may even go so far as employing speech-recognition software that registers strings of keywords or facial recognition software that identifies known terrorists



=======================================================


Fast  Forward  {  year  2020  } :


Sergy Brin / Elon Musk / Satya Nadela / Mark Zuckerberg / Jeff Bezos / Ray Kurzweil / Tim Cook / Vishal Sikka / Nandan Nilekani / Pramod Verma etc , come together and examine the following  :





==================================================




==================================================





====================================================



======================================================


04  Aug  2017



No comments:

Post a Comment