Tech Companies Move to Target Terrorist Propaganda Online | NBC 6 South Florida
National & International News
The day’s top national and international news

Tech Companies Move to Target Terrorist Propaganda Online

The companies promised to share among themselves "the most extreme and egregious terrorist images and videos we have removed from our services"



    Getty Images/Science Photo Library (File)
    Secure website.

    Facebook, Microsoft, Twitter and YouTube are joining forces to more quickly identify the worst terrorist propaganda and prevent it from spreading online.

    The new program announced Monday would create a database of unique digital "fingerprints" to help automatically identify videos or images the companies could remove.

    Gov't Agencies to Guard Against Election Day Cyber Attacks

    [NATL] Government Agencies to Guard Against Election Day Cyber Attacks
    U.S. Intelligence sources have confirmed they're preparing for possible hacks affecting electricity, the internet, social media and transportation next Tuesday.
    (Published Friday, Nov. 4, 2016)

    The move by the technology companies, which is expected to begin in early 2017, aims to assuage government concerns — and derail proposed new federal legislation — over social media content that is seen as increasingly driving terrorist recruitment and radicalization, while also balancing free-speech issues. 

    Technical details were being worked out, but Microsoft pioneered similar technology to detect, report and remove child pornography through such a database in 2009. Unlike those images, which are plainly illegal under U.S. law, questions about whether an image or video promotes terrorism can be more subjective, depending on national laws and the rules of a particular company's service.

    Social media has increasingly become a tool for recruiting and radicalization by the Islamic State group and others. Its use by terror groups and supporters has added to the threat from so-called lone-wolf attacks and decreased the time from "flash to bang" — or radicalization to violence — with little or no time for law enforcement to follow evidentiary trails before an attack.

    Under the new partnership, the companies promised to share among themselves "the most extreme and egregious terrorist images and videos we have removed from our services — content most likely to violate all our respective companies' content policies," according to a joint announcement Monday evening.

    When such content is shared internally, the other participating companies will be notified and can use the digital fingerprints to quickly identify the same content on their own services to judge whether it violates their rules. If so, companies can delete the material and possibly disable the account, as appropriate. 

    Most social media services explicitly do not allow content that supports violent action or illegal activities. Twitter, for example, says users "may not promote violence against or directly attack or threaten other people on the basis of race, ethnicity, national origin, sexual orientation, gender, gender identity, religious affiliation, age, disability or disease."

    "We really are going after the most obvious serious content that is shared online — that is, the kind of recruitment videos and beheading videos more likely to be against all our content policies," said Sally Aldous, a Facebook spokeswoman.

    The new program caps a year of efforts to tamp down on social media's use by terrorist groups.

    Flight Attendant's Behavior Questioned Following Altercation

    [NATL-BAY] Flight Attendant's Behavior Questioned Following Altercation

    One day after an American Airlines flight attendant became entangled in an altercation with passengers, his employer is investigating whether or not he was out of line. Thom Jensen reports.

    (Published Sunday, April 23, 2017)

    Lawmakers last year introduced legislation that would require social media companies to report any online terrorist activity they became aware of to law enforcement. The bill by Sens. Dianne Feinstein, D-Calif., and Richard Burr, R-N.C., was criticized for not defining "terrorist activity," which could have drowned government agencies in reports. The bill was opposed by the Internet Association, which represents 37 internet companies, including Facebook, Snapchat, Google, LinkedIn, Reddit, Twitter, Yahoo and others.

    The bill came days after Syed Farook and his wife, Tashfeen Malik, went on a shooting attack in San Bernardino, California, killing 14 people and injuring 21 others. A Facebook post on Malik's page around the time of the attack included a pledge of allegiance to the leader of the Islamic State group.

    Facebook found the post — which was under an alias — the day after the attack. The company removed the profile from public view and informed law enforcement. Such a proactive effort had previously been uncommon.

    Twitter moved toward partial automation in late 2015, using unspecified "proprietary spam-fighting tools" to find accounts that might be violating its terms of service and promoting terrorism. The material still required review by a team at Twitter before the accounts could be disabled.

    Drive-Thru Pot Dispensary Opens in Colorado

    [NATL] Drive-Thru Pot Dispensary Opens in Colorado

    When you think of a drive-thru, fast food probably comes to mind. But what about marijuana? Tumbleweed Express Drive-Thru in Parachute, Colorado, is the first of its kind. After nearly a year of preparation, the drive-thru pot shop opened for business on 4/20. 

    "Their timing couldn't have been better in my opinion, I mean everybody wants to celebrate," said Kyle Steele, a customer at the drive-thru.

    (Published Friday, April 21, 2017)

    "Since the middle of 2015, we have suspended more than 360,000 accounts for violating Twitter's policy on violent threats and the promotion of terrorism," said Sinead McSweeney, Twitter's vice president of public policy. "A large proportion of these accounts have been removed by technical means, including our proprietary spam-fighting tools."

    Facebook has also used image-matching technology to compare images to ones it's already removed. The effort lets Facebook review images to avoid removing legitimate and protected uses, such as a photograph published by a news organization, a spokeswoman said.

    Terrence McNeil of Ohio was charged in 2015 with soliciting the killings of U.S. service members over social media, including Tumblr, Facebook and Twitter. Federal prosecutors accused him of posting a series of photographs on his Facebook account to praise the death of a Jordanian pilot who was burned to death by the Islamic State group — showing him before, during and after his death, including an image of him engulfed in flames, according to the complaint. 

    In January, the White House dispatched top officials, including FBI Director James Comey, Attorney General Loretta Lynch and National Security Agency Director Mike Rogers, to Silicon Valley to discuss the use of social media by violent extremist groups. Among the issues they discussed was how to use technology to help quickly identify terrorist content.

    Protester, Official React to Arkansas Execution

    [NATL] Protester, Official React to Arkansas Execution

    Just before midnight Thursday, Arkansas announced it had executed death row inmate Ledell Lee. He is the first of four inmates scheduled to die before the end of the month when a crucial lethal injection drug is set to expire. A protester and a spokesperson for the governor reacted to the controversial execution.

    (Published Friday, April 21, 2017)

    The four companies say they will be looking at involving additional companies in the future.