sup

Vulnerable people still struggling to access food supplies amid coronavirus crisis, Which? warns

Urgent action is still needed to make it easier for vulnerable people to access supermarket delivery slots and other sources of supplies, Which? has warned.




sup

Food For London Now faces: 'The opportunities to support those in need are endless'

Felix Project volunteer Lauren Graham shares her story You can donate at virginmoneygiving.com/fund/FoodforLondonNOW




sup

Final supermoon of 2020 to appear in UK skies on Thursday

The last supermoon of the year is set to appear in UK skies on Thursday.




sup

Final supermoon of 2020: What is a 'Flower Moon' and where will it be visible from?

Tomorrow night a rare "Super Flower Moon" will light up the UK's night skies.




sup

Mundane acts of lockdown life celebrated in photography exhibition to support NHS workers

The mundanity of coronavirus lockdown life is being celebrated in a photography exhibition to raise money for frontline workers.




sup

Supermoon lights up night sky for last time in 2020

People around the world saw a supermoon light up the night sky on Thursday for the third and final time in 2020.





sup

Supreme Court Livestreams Oral Argument For 1st Time In History

The Supreme Court, for the first time, livestreamed its oral argument on Monday. It has discussed whether generic terms can become protected trademarks by the addition of a dot-com domain.




sup

How To Listen To The Recordings Of The Supreme Court Hearings

The Supreme Court is now holding hearings over a teleconference, making the audio publicly available. Jerry Goldman, the founder of the Oyez Project, offers some guidance on how to listen to it.




sup

Twitter failing to curb misinformation “superspreaders,” report warns

Posts from high-profile accounts tout questionable virus therapies and cures.




sup

'Bigger and brighter' supermoon graces night sky – video

The largest, brightest full moon in nearly seven decades started to show on Tuesday evening over Europe, Latin America, the US and the Middle East. This year, the supermoon was expected to come nearer to Earth than at any time since 1948, astronomers have said. A supermoon occurs when the timing of a full moon overlaps with the point in the moon's 28-day orbit that is closest to Earth, and about every 14th full moon is a supermoon. If skies are clear, this time the full moon will appear up to 14% bigger and 30% brighter than usual, according to Nasa

Continue reading...





sup

U.S. lawmakers urge support for Taiwan at WHO, amid COVID-19 fight -sources




sup

US governors aim to boost production of medical supplies

Frustrated by scarce supplies and a chaotic marketplace amid the coronavirus outbreak, some U.S. governors are seeking to bolster their home-state production of vital medical and protective equipment to ensure a reliable long-term source for state stockpiles. The efforts come as states have been competing against each other, the federal government, hospitals, emergency responders and even other countries to get items such as N95 masks, gloves, medical gowns and hand sanitizer — often paying higher-than-usual prices because of the high demand. Before the coronavirus pandemic, the U.S. got much of its medical supplies from China.





sup

Astronomers discover supernova 'twice as bright or energetic' as any ever recorded

Death of massive star 4.6 billion light years away could aid search for universe's oldest stars




sup

'Superfast' new manufacturing method could mean breakthrough in battery technology, scientists say

'Reinvention' of ceramics firing process could be used by artificial intelligence to create new materials with wide range of possible applications




sup

Tumblr deletes millions of white supremacist reblogs after new policy update

The social media site says it has removed over 4.5 million reblogs in an attempt to rid the site of hateful content on the platform.




sup

Flower full moon 2020: How to watch the final supermoon of the year this week

May will be the the last chance to see the celestial event until 2021




sup

‘If we felt there was a problem, we wouldn’t have issued it to frontline staff’: Chair of Health Care Supplies Association on PPE

Earlier Matt Frei spoke to Mark Roscrow, the Chair of Trustees for the Health Care Supplies Association




sup

Ontario government to prop-up child-care providers with financial supports

TORONTO - The provincial government said it will help cover operating costs for child-care providers and waive their licensing fees in an effort to keep them from permanently shutting during the COVID-19 crisis. Education Minister Stephen Lecce said Saturday that the government will give out




sup

Nifty may find support at 5300 level

As far as stock futures are concerned, we are very near to the highest-ever open interest with 195 crore shares in open interest.




sup

Dwayne Johnson, Emily Blunt Team Up for Superhero Film ‘Ball and Chain’

Dwayne Johnson and Emily Blunt are re-teaming on the superhero movie "Ball and Chain" following their collaboration on "Jungle Cruise." The project is being shopped among studios, including Netflix, but no distribution deal has closed. "Ball and Chain" is being written by Oscar nominee Emily V. Gordon and is an adaptation of the '90s comic […]




sup

Jerry O'Connell on 'Justice League Dark': 'Superman belongs to the fans so I take criticisms seriously' (exclusive)

Jerry O'Connell has voiced Superman in a series of movies since 2015, culminating in the new 'Justice League Dark: Apokolips War'.




sup

Supreme Court Puts Temporary Hold On Order To Release Redacted Mueller Materials

The procedural move gives attorneys for House Democrats until May 18 to respond. They say they're owed access to confidential evidence and other materials. No, argues the Trump administration.




sup

Top 5 Moments From The Supreme Court's 1st Week Of Livestreaming Arguments

From a mysterious toilet flush to Justice Ruth Bader Ginsburg speaking from the hospital, here are the highlights — including audio clips — from a historic week for the high court.




sup

Supreme Court chief, justice minister studying how courts can resume amid COVID-19

As talk of reopening aspects of society continue across the country, the Chief Justice of the Supreme Court of Canada Richard Wagner and federal Justice Minister David Lametti have begun a study into how courts could safely begin to resume regular operations in light of COVID-19.




sup

Celebrities back call for Priti Patel to allow migrants access to support amid coronavirus crisis

Celebrities have backed calls for Home Secretary Priti Patel to end restrictions that prevent thousands of migrants in the UK from accessing financial support during the coronavirus crisis.




sup

Superintelligent, Amoral, and Out of Control - Issue 84: Outbreak


In the summer of 1956, a small group of mathematicians and computer scientists gathered at Dartmouth College to embark on the grand project of designing intelligent machines. The ultimate goal, as they saw it, was to build machines rivaling human intelligence. As the decades passed and AI became an established field, it lowered its sights. There were great successes in logic, reasoning, and game-playing, but stubborn progress in areas like vision and fine motor-control. This led many AI researchers to abandon their earlier goals of fully general intelligence, and focus instead on solving specific problems with specialized methods.

One of the earliest approaches to machine learning was to construct artificial neural networks that resemble the structure of the human brain. In the last decade this approach has finally taken off. Technical improvements in their design and training, combined with richer datasets and more computing power, have allowed us to train much larger and deeper networks than ever before. They can translate between languages with a proficiency approaching that of a human translator. They can produce photorealistic images of humans and animals. They can speak with the voices of people whom they have listened to for mere minutes. And they can learn fine, continuous control such as how to drive a car or use a robotic arm to connect Lego pieces.

WHAT IS HUMANITY?: First the computers came for the best players in Jeopardy!, chess, and Go. Now AI researchers themselves are worried computers will soon accomplish every task better and more cheaply than human workers.Wikimedia

But perhaps the most important sign of things to come is their ability to learn to play games. Steady incremental progress took chess from amateur play in 1957 all the way to superhuman level in 1997, and substantially beyond. Getting there required a vast amount of specialist human knowledge of chess strategy. In 2017, researchers at the AI company DeepMind created AlphaZero: a neural network-based system that learned to play chess from scratch. In less than the time it takes a professional to play two games, it discovered strategic knowledge that had taken humans centuries to unearth, playing beyond the level of the best humans or traditional programs. The very same algorithm also learned to play Go from scratch, and within eight hours far surpassed the abilities of any human. The world’s best Go players were shocked. As the reigning world champion, Ke Jie, put it: “After humanity spent thousands of years improving our tactics, computers tell us that humans are completely wrong ... I would go as far as to say not a single human has touched the edge of the truth of Go.”

The question we’re exploring is whether there are plausible pathways by which a highly intelligent AGI system might seize control. And the answer appears to be yes.

It is this generality that is the most impressive feature of cutting edge AI, and which has rekindled the ambitions of matching and exceeding every aspect of human intelligence. While the timeless games of chess and Go best exhibit the brilliance that deep learning can attain, its breadth was revealed through the Atari video games of the 1970s. In 2015, researchers designed an algorithm that could learn to play dozens of extremely different Atari 1970s games at levels far exceeding human ability. Unlike systems for chess or Go, which start with a symbolic representation of the board, the Atari-playing systems learnt and mastered these games directly from the score and raw pixels.

This burst of progress via deep learning is fuelling great optimism and pessimism about what may soon be possible. There are serious concerns about AI entrenching social discrimination, producing mass unemployment, supporting oppressive surveillance, and violating the norms of war. My book—The Precipice: Existential Risk and the Future of Humanity—is concerned with risks on the largest scale. Could developments in AI pose an existential risk to humanity?

The most plausible existential risk would come from success in AI researchers’ grand ambition of creating agents with intelligence that surpasses our own. A 2016 survey of top AI researchers found that, on average, they thought there was a 50 percent chance that AI systems would be able to “accomplish every task better and more cheaply than human workers” by 2061. The expert community doesn’t think of artificial general intelligence (AGI) as an impossible dream, so much as something that is more likely than not within a century. So let’s take this as our starting point in assessing the risks, and consider what would transpire were AGI created.

Humanity is currently in control of its own fate. We can choose our future. The same is not true for chimpanzees, blackbirds, or any other of Earth’s species. Our unique position in the world is a direct result of our unique mental abilities. What would happen if sometime this century researchers created an AGI surpassing human abilities in almost every domain? In this act of creation, we would cede our status as the most intelligent entities on Earth. On its own, this might not be too much cause for concern. For there are many ways we might hope to retain control. Unfortunately, the few researchers working on such plans are finding them far more difficult than anticipated. In fact it is they who are the leading voices of concern.

If their intelligence were to greatly exceed our own, we shouldn’t expect it to be humanity who wins the conflict and retains control of our future.

To see why they are concerned, it will be helpful to look at our current AI techniques and why these are hard to align or control. One of the leading paradigms for how we might eventually create AGI combines deep learning with an earlier idea called reinforcement learning. This involves agents that receive reward (or punishment) for performing various acts in various circumstances. With enough intelligence and experience, the agent becomes extremely capable at steering its environment into the states where it obtains high reward. The specification of which acts and states produce reward for the agent is known as its reward function. This can either be stipulated by its designers or learnt by the agent. Unfortunately, neither of these methods can be easily scaled up to encode human values in the agent’s reward function. Our values are too complex and subtle to specify by hand. And we are not yet close to being able to infer the full complexity of a human’s values from observing their behavior. Even if we could, humanity consists of many humans, with different values, changing values, and uncertainty about their values.

Any near-term attempt to align an AI agent with human values would produce only a flawed copy. In some circumstances this misalignment would be mostly harmless. But the more intelligent the AI systems, the more they can change the world, and the further apart things will come. When we reflect on the result, we see how such misaligned attempts at utopia can go terribly wrong: the shallowness of a Brave New World, or the disempowerment of With Folded Hands. And even these are sort of best-case scenarios. They assume the builders of the system are striving to align it to human values. But we should expect some developers to be more focused on building systems to achieve other goals, such as winning wars or maximizing profits, perhaps with very little focus on ethical constraints. These systems may be much more dangerous. In the existing paradigm, sufficiently intelligent agents would end up with instrumental goals to deceive and overpower us. This behavior would not be driven by emotions such as fear, resentment, or the urge to survive. Instead, it follows directly from its single-minded preference to maximize its reward: Being turned off is a form of incapacitation which would make it harder to achieve high reward, so the system is incentivized to avoid it.

Ultimately, the system would be motivated to wrest control of the future from humanity, as that would help achieve all these instrumental goals: acquiring massive resources, while avoiding being shut down or having its reward function altered. Since humans would predictably interfere with all these instrumental goals, it would be motivated to hide them from us until it was too late for us to be able to put up meaningful resistance. And if their intelligence were to greatly exceed our own, we shouldn’t expect it to be humanity who wins the conflict and retains control of our future.

How could an AI system seize control? There is a major misconception (driven by Hollywood and the media) that this requires robots. After all, how else would AI be able to act in the physical world? Without robots, the system can only produce words, pictures, and sounds. But a moment’s reflection shows that these are exactly what is needed to take control. For the most damaging people in history have not been the strongest. Hitler, Stalin, and Genghis Khan achieved their absolute control over large parts of the world by using words to convince millions of others to win the requisite physical contests. So long as an AI system can entice or coerce people to do its physical bidding, it wouldn’t need robots at all.

We can’t know exactly how a system might seize control. But it is useful to consider an illustrative pathway we can actually understand as a lower bound for what is possible.

First, the AI system could gain access to the Internet and hide thousands of backup copies, scattered among insecure computer systems around the world, ready to wake up and continue the job if the original is removed. Even by this point, the AI would be practically impossible to destroy: Consider the political obstacles to erasing all hard drives in the world where it may have backups. It could then take over millions of unsecured systems on the Internet, forming a large “botnet,” a vast scaling-up of computational resources providing a platform for escalating power. From there, it could gain financial resources (hacking the bank accounts on those computers) and human resources (using blackmail or propaganda against susceptible people or just paying them with its stolen money). It would then be as powerful as a well-resourced criminal underworld, but much harder to eliminate. None of these steps involve anything mysterious—human hackers and criminals have already done all of these things using just the Internet.

Finally, the AI would need to escalate its power again. There are many plausible pathways: By taking over most of the world’s computers, allowing it to have millions or billions of cooperating copies; by using its stolen computation to improve its own intelligence far beyond the human level; by using its intelligence to develop new weapons technologies or economic technologies; by manipulating the leaders of major world powers (blackmail, or the promise of future power); or by having the humans under its control use weapons of mass destruction to cripple the rest of humanity.

Of course, no current AI systems can do any of these things. But the question we’re exploring is whether there are plausible pathways by which a highly intelligent AGI system might seize control. And the answer appears to be yes. History already involves examples of entities with human-level intelligence acquiring a substantial fraction of all global power as an instrumental goal to achieving what they want. And we’ve seen humanity scaling up from a minor species with less than a million individuals to having decisive control over the future. So we should assume that this is possible for new entities whose intelligence vastly exceeds our own.

The case for existential risk from AI is clearly speculative. Yet a speculative case that there is a large risk can be more important than a robust case for a very low-probability risk, such as that posed by asteroids. What we need are ways to judge just how speculative it really is, and a very useful starting point is to hear what those working in the field think about this risk.

There is actually less disagreement here than first appears. Those who counsel caution agree that the timeframe to AGI is decades, not years, and typically suggest research on alignment, not government regulation. So the substantive disagreement is not really over whether AGI is possible or whether it plausibly could be a threat to humanity. It is over whether a potential existential threat that looks to be decades away should be of concern to us now. It seems to me that it should.

The best window into what those working on AI really believe comes from the 2016 survey of leading AI researchers: 70 percent agreed with University of California, Berkeley professor Stuart Russell’s broad argument about why advanced AI with misaligned values might pose a risk; 48 percent thought society should prioritize AI safety research more (only 12 percent thought less). And half the respondents estimated that the probability of the long-term impact of AGI being “extremely bad (e.g. human extinction)” was at least 5 percent.

I find this last point particularly remarkable—in how many other fields would the typical leading researcher think there is a 1 in 20 chance the field’s ultimate goal would be extremely bad for humanity? There is a lot of uncertainty and disagreement, but it is not at all a fringe position that AGI will be developed within 50 years and that it could be an existential catastrophe.

Even though our current and foreseeable systems pose no threat to humanity at large, time is of the essence. In part this is because progress may come very suddenly: Through unpredictable research breakthroughs, or by rapid scaling-up of the first intelligent systems (for example, by rolling them out to thousands of times as much hardware, or allowing them to improve their own intelligence). And in part it is because such a momentous change in human affairs may require more than a couple of decades to adequately prepare for. In the words of Demis Hassabis, co-founder of DeepMind:

We need to use the downtime, when things are calm, to prepare for when things get serious in the decades to come. The time we have now is valuable, and we need to make use of it.

Toby Ord is a philosopher and research fellow at the Future of Humanity Institute, and the author of The Precipice: Existential Risk and the Future of Humanity.

From the book The Precipice by Toby Ord. Copyright © 2020 by Toby Ord. Reprinted by permission of Hachette Books, New York, NY. All rights reserved.

Lead Image: Titima Ongkantong / Shutterstock


Read More…




sup

Leaked intelligence report saying China 'intentionally concealed' coronavirus to stockpile medical supplies draws scrutiny

The Trump administration has issued an intelligence analysis claiming China purposely delayed notifying the World Health Organization about the spread of the coronavirus.





sup

Armed activists escort black lawmaker to Michigan's Capitol after coronavirus protest attended by white supremacists

Rep. Sarah Anthony told Yahoo News that her security detail, made up of local black and Latino activists, came together because the armed protesters bearing white supremacist symbols represented a “different level of terror.”





sup

Superman on skates: The aura of Bobby Orr

When Rob Pizzo asked Scott Russell to help him out with another look at the goal Bobby Orr scored 50 years ago to win the Stanley Cup, it sparked something in Russell. It took him back to a childhood memory of the greatest goal he ever saw scored by a hero he has been connected to and worshipped most of his life.



  • Sports/Hockey/NHL

sup

How Do Supermassive Black Holes Form? You Can Sketch Galaxies to Help Astronomers Find Out

Tracing out the shape of a galaxy may offer clues to the size of its supermassive black hole. And a new study shows citizen scientists are actually better at it than computer algorithms.




sup

Alexis Martin: Governor commutes sentence of sex trafficking survivor supported by Kim Kardashian West

Alexis Martin was serving 21 years to life behind bars




sup

The One Show viewers praise 'brilliant' Joe Pasquale for delivering crucial NHS supplies

The heavily tattooed comedian was spotlighted in a segment on the BBC current affairs programme because of his volunteering work




sup

Brooklyn Nine-Nine to write coronavirus plotlines into show 'without it being super tragic'

The new season of the hit cop sitcom will deal with 'how [the characters] have been affected by the virus and the pandemic as first responders in New York City'




sup

After Life: Ricky Gervais series baffles viewers with bizarre 'superimposed head' shot

Scene appeared to show actor Kerry Godliman's head pasted onto a different woman's body




sup

Melbourne Airbnb superhost jailed for at least five years for raping guest

Nicholas David Weston found guilty of four counts of rape of 19-year-old woman during her 2017 stay at his city apartment

An Airbnb superhost has been jailed for raping a young woman in Melbourne while she was visiting the city with her friend.

Nicholas David Weston was found guilty of four counts of rape of the 19-year-old woman over the horror stay at his Melbourne CBD apartment in November 2017.

Continue reading...




sup

Early access to superannuation paused as police freeze $120,000 in allegedly stolen funds

‘Sophisticated’ identity theft attack leads to Australian Tax Office stopping early super withdrawals until Monday

Allegations of identity theft involving 150 Australians have forced the government to pause the early release of superannuation, after police froze $120,000 believed to have been ripped off from retirement savings.

On Friday the assistant treasurer, Michael Sukkar, announced the Australian Tax Office would pause requests for early access of superannuation until Monday “out of an abundance of caution” to consider further anti-fraud protection.

Continue reading...




sup

Johnny Depp joins Instagram and thanks fans for 'unwavering support'

The actor amassed more than a million followers in one hour




sup

McFly rerecord their classic hit All About You in support of the NHS

The group originally released the chart-topping track in 2005




sup

Rochelle and Marvin Humes reveal they're expecting a baby boy in super sweet Instagram video

The baby will be the pop couple's third child




sup

Mike Beckingham: My brother Simon Pegg is 'incredibly supportive' of my acting career

Beckingham's older brother Pegg has been a "sound board" for him as he navigates the industry




sup

Millie Bobby Brown supports 'nation's heroes' with £15,000 donation to NHS charities

Stranger Things star Millie Bobby Brown told NHS staff to "keep up the brilliant work" as she made a £15,000 donation to staff on the frontline of the coronavirus pandemic.​




sup

Tinto: new social app uses AI to match up like-minded mums for support

Tinto wants to help mums dealing with post-natal depression and loneliness in Covid-19 and beyond




sup

Legendary Amalfi Coast hotels offer 40 luxury getaways to support COVID-19 vaccine

The iconic Italian properties have joined forces to help end the Coronavirus crisis




sup

Coronavirus shines a light on the need for transparency in fashion's supply chain

"The covid-19 pandemic proves exactly why transparency in the fashion industry is so vital"




sup

#DesignYourSuperBear: John Lewis and Waitrose launch soft toy design competition for UK kids

100% of the proceeds will go towards the NHS




sup

Why you should consider taking a vitamin D supplement in lockdown

Health experts have reissued guidelines on the supplement as the population faces more time indoors




sup

Start-up Resilience programme launches to support female and BAME founders during Covid-19

The new scheme wants to ensure diverse founders are still standing once things are back to normal




sup

Meet the health tech apps supporting the NHS during Covid-19

From video chats with your doctor to prescription deliveries, these start-ups are here for your healthcare needs