当赵丽颖谭松韵娄艺潇老了…应当是这样的吧?_娱乐频道_凤凰网霍

2018-03-17 10:54

都说美是女人终生寻求的事业,有事没事都敷个面膜,天天和朽迈作奋斗!

但是当年华逝去,现在大家爱好的爱豆又会变成什么样子呢?

有网友在网上找出来了杨昆和赵丽颖的古装对比图,相似的脸型和轮廓,确切是让人感慨一句,这个预测还是有一定可托度的!(图源八组)

赵丽颖和杨昆的大眼睛、圆圆脸异常相似,

小妹又找了两张图做对比,虽然一张古装、一张现代装,但是角度和发型、眉毛都有些相像之处。虽然杨昆现在演得都是一些婆婆妈妈,但美人还是丽人,果然是“长得难看的人都差未几,而丑人却各有各的丑法;,或者赵丽颖老了以后就是这样吧?

小妹忽然想感叹一句,岁月不饶人啊,大家必定要留神保养!

谭松韵和《神医喜来乐》里面的梁丽,两个人的眼睛和脸型很相似,大略谭松韵老了当前会变成这样!(图源八组)

小妹也去做了2张比较图,大家来感触一下,虽然是时装和古代装的对照,然而眉型还是很像的。

下巴可以说十分类似了!

看完了这组对比,还有娄艺潇和杨青,网友拼出来两个人的图是这样的,小妹感到主要还是下垂眼和脸型比拟像……

不外这一切都建立在这些年青的女明星没有往脸上emm……对吧?看到这里,小妹我也信心要好好颐养了!


原题目:霍金演讲:人工智能也可能是人类文明的终结者

4月27日,有名物理学家史蒂芬·霍金在北京举行的全球挪动互联网大会上做了视频演讲。在报告中,霍金重申人工智能突起要么是人类最好的事情,要么就是最蹩脚的事情。他认为,人类需警戒人工智能发展要挟。因为人工智能一旦脱离束缚,以不断加速的状态重新设计自身,人类因为受到漫长的生物进化的制约,将无法与之竞争,从而被取代。

霍金演讲稿双语全文&采访对话稿

一、人工智能,要么是最好的,要么是最糟的

在我的一生中,我见证了许多社会深刻的变化。其中最深入,同时也是对人类影响一劳永逸的变化就是人工智能的崛起。

简略来说,我认为壮大的人工智能的崛起,要么是人类历史上最好的事,要么是最糟的。

是好是坏,我不得不说我们依然不能确定,3494天线宝宝最快开奖。但我们应该竭尽所能,确保其未来发展对我们的后辈和环境有利。

咱们别无抉择。我以为人工智能的发展,自身是一种存在着问题的趋势,而这些问题必需在当初跟未来得到解决。

人工智能的研究与开发正在迅速推动。也许科学研究应该暂停片刻,从而使研究重点从晋升人工智能能力转移到最大化人工智能的社会效益上面。

基于这样的斟酌,美国人工智能协会(AAAI)于2008至2009年,成立了人工智能长期未来总筹论坛。他们近期在目的导向的中性技术上投入了大量的关注。但人工智能系统的原则仍然必须要依照我们的意志工作。

跨学科研究可能是一条可能的前进途径:从经济、法律、哲学延长至计算机保险、形式化方式,当然还有人工智能本身的各个分支。

文化所提发生的所有都是人类智能的产物,我信任生物大脑总有一天会达到计算机可以到达的水平,不实质差别。因而,它遵守了“盘算机在实践上能够模拟人类智能,而后超出;这一准则。

但我们并不肯定,所以我们无法晓得我们将无穷地得到人工智能的赞助,还是被鄙弃并被边沿化,或者很可能被它覆灭。确实,我们担忧聪慧的机器将可能取代人类正在从事的工作,并迅速地毁灭数以百万计的工作岗位。

在人工智能从原始形态不断发展,并被证实无比有用的同时,我也在担心这样这个成果,即创造一个可以等同或超越人类的智能的人工智能:人工智能一旦脱离约束,以一直加速的状况从新设计本身。

人类因为受到漫长的生物进化的限度,无法与之竞争,将被代替。这将给我们的经济带来极大的损坏。未来,人工智能可以发展出自我意志,一个与我们抵触的意志。

良多人认为人类可以在相称长的时间里把持技巧的发展,这样我们就能看到人工智能可以解决世界上大部门问题的潜力。但我并不断定,只管我对人类一贯持有乐观的立场。

二、人工智能对社会所造成的影响,需要认真调研

2015年1月份,我和科技企业家埃隆·马斯克,以及很多其余的人工智能专家签订了一份关于人工智能的公开信。目标是倡导就人工智能对社会所造成的影响,做当真的调研。

在这之前,埃隆·马斯克就忠告过人们:超人类人工智能可能带来不可估计的好处。但如果安排不当,则可能给人类带来相反的后果。我和他同在“性命未来研究所;担负科学参谋委员会的职务,这是一个旨在缓解人类所面临的存在风险的组织。

之前提到的公开信也是由这个组织起草的。这个公开信号令开展可以禁止潜在问题的直接研究,同时也播种人工智能带给我们的潜在利益,同时致力于让人工智能的研发人员更关注人工智能平安。

对于决议者和一般民众来说,这封公开信内容翔实,并非危言耸听。人人都知道人工智能,我们认为这一点非常重要。比如,人工智能存在铲除疾患和贫苦的潜力,但是研究人员必须能够保障创造出可控的人工智能。

那封只有四段文字,标题为《应优先研究强盛而有利的人工智能》的公然信,在其附带的十二页文件中对研讨的优先顺序作了具体的部署。

在过去的20年或更长时间里,人工智能一直专一于建设智能代办所产生的问题,即:在特定环境下可以感知并行为的各种系统。

智能是一个与统计学和经济学相干的感性概念。艰深地讲,这是一种能做出好的决议、打算和推论的才能。基于这些工作,大量的整合和穿插孕育被利用在人工智能、机器学习、统计学、节制论、神经科学以及其它领域。

共享理论框架的建立,联合数据的供给和处置能力,在各种细分的领域取得了明显的成功。

例如语音辨认、图像分类、主动驾驶、机器翻译、步态运动和问答系统。

随着这些领域的发展,从试验室研究到有经济价值的技术构成了良性轮回。哪怕很小的机能改良,都会带来巨大的经济效益,进而激励更长期、更巨大的投入和研究。目前人们广泛认同,人工智能的研究正在稳步发展,而它对社会的影响很可能还在扩展。

潜在的利益是宏大的,甚至文明所产生的一切,都可能是人类智能的产物。但我们无奈猜测我们会获得什么结果,这种成果可能是被人工智能工具放大过的。

正如我说的,根除疾病和贫困并不是完全不可能,由于人工智能的巨大潜力,研究如何(从人工智能)获益并躲避风险长短常重要的。

三、从短期和长期看人工智能

现在,关于人工智能的研究正在敏捷发展,这一研究可以从短期和长期来分辨探讨。

短期担忧:

1.无人驾驶。

从民用无人机到自主驾驶汽车。在紧迫情况下,一辆无人驾驶汽车不得不在小风险的大事故和或许率的小事变之间进行取舍。

2.致命性智能自主武器。

它们是否该被制止?如果是,那么“自主;该如何准确定义。如果不是,任何应用不当和故障的差错应该如何被问责。

3.隐衷的担忧。
 

由于人工智能逐渐开端解读大量监控数据,会造成隐私上的担忧,以及如何治理因人工智能取代工作岗位带来的经济影响。

长期担忧重要是人工智能体系失控的潜在风险。

随着不遵循人类志愿行事的超级智能的崛起,强大的系统可能会威逼到人类发展。这种错位是否会产生?如果会,那些情形是如何呈现的?我们应该投入什么样的研究,以便更好的理解和解决危险的超级智能崛起的可能性,或智能暴发的涌现?

当前掌握人工智能技术的工具,例如强化学习,简单适用的功效,还不足以解决这系统失控的问题。因此,我们需要进一步研究来找到和确认一个牢靠的解决措施来掌控这一问题。

近来的里程碑,比方之条件到的自主驾驶汽车,以及人工智能博得围棋竞赛,都是未来趋势的迹象,伟大的投入倾泻到这项科技。我们目前所取得的成就,和未来几十年后可能取得的成就比拟必定黯然失色。

而且我们远不能预测我们能取得什么成就。当我们的脑筋被人工智能放大以后,也许在这种新技术革命的帮助下,我们可以解决一些工业化对天然界造成的侵害。关乎到我们生活的各个方面都行将被改变。

简而言之,人工智能的胜利有可能是人类文明史上最大的事件。

但人工智能也有可能是人类文明史的终结,除非我们学会如何防止危险。我曾经说过,人工智能的全方位发展可能导致人类的消亡。好比最大化使用智能性自主武器。今年早些时候,我和一些来自世界各国的科学家独特在结合国会议上支撑其对核武器的禁令。我们正在着急的等候协商结果。

目前,九个核大国可以控制大概一万四千个核武器,它们中的任何一个国度都可以将城市夷为平川,喷射性废料会大面积传染农田,最可怕的迫害是诱发核冬天,火和烟雾会导致全球的小冰河期。

这一结果使寰球食粮系统崩塌,末日般动荡,很可能导致大部分人逝世亡。我们作为科学家,对核兵器承当着特别的义务。恰是科学家发现了核武器,并发明它们的影响比最初料想的更加恐怖。

我对人工智能的灾难探讨可能惊吓到了各位。很负疚。但是作为今天的与会者,重要的是,在影响当前技术的未来研发中,你们要明白本人所处的地位。

我相信我们团结在一起,来呼吁国际公约的支持或者签署呈交给各国政府的公开信,科技首领和科学家正极尽所能避免不可控的人工智能的崛起。

去年10月,我在英国剑桥建立了一个新的机构,试图解决一些在人工智能研究疾速发展中出现的尚无定论的问题。“利弗休姆智能未来中心;是一个跨学科研究所,致力于研究智能的未来,这对我们文明和物种的未来至关重要。

从前我们破费大批时光学习历史,固然深刻去看,可能大多数是关于笨拙的历史。所以现在人们转而研究智能的将来,是令人惊喜的变更。

我们对潜在危险有所意识,我心坎仍秉持乐观态度,我相信创造智能的潜在收益是巨大的。也许借助这项新技术革命的工具,我们将可以削减工业化对做作界造成的损害。

我们生涯的每一个方面都会被转变。我在研究所的共事休·普林斯否认,“利弗休姆核心;能树立,局部是由于大学成破了“存量危险中央;。后者更加普遍地审阅人类潜在问题,而“利弗休姆中央;的重点研究范畴则绝对狭小。

四、人工智能的最新进展

人工智能的最新进展,包括欧洲议会呼吁起草一系列法规,以管理机器人和人工智能的创新。令人些许惊奇的是,这里涉及到了一种形式的电子人格。以确保最有能力和最进步的人工智能尽到其应尽的权利和责任。

欧洲议会发言人评论说,跟着日常生活中越来越多的范畴日益受到机器人的影响,我们须要确保机器人无论现在仍是将来,都为人类而服务。

向欧洲议会议员提交的讲演,明白认为世界正处于新的产业机器人革命的前沿。呈文中剖析了是否应该给机器人供给作为电子人的权力。这等同于法人(的身份),兴许有可能。

报告强调,在任何时候,研究和设计职员都应确保每一个机器人设计都包括有终止开关。在库布里克的片子《2001太空周游》中,出故障的超级电脑哈尔没有让科学家们进入太空舱,但那是科幻。我们要面对的则是事实。

奥斯本·克拉克跨国律师事务所的合伙人,洛纳·布拉泽尔在报告中说,我们不承认鲸鱼和大猩猩有人格,所以也没有必要急于接收一个机器人人格。但是担忧始终存在。

报告承认在多少十年的时间内,人工智能可能会超越人类智力范围,人工智能可能会超越人类智力规模,进而挑衅人机关联。

报告最后呐喊成立欧洲机器人和人工智能机构,以提供技术、伦理和监管方面的专业常识。如果欧洲议会议员投票同意立法,该报告将提交给欧盟委员会。它将在三个月的时间内决定要采取哪些立法步骤。

在人工智能发展过程中,我们还应该表演一个角色,确保下一代不仅仅有机遇还要有决心。在早期阶段充足参加迷信研究,以便他们持续施展潜力,辅助人类发明一个更加美妙的的世界。这就是我刚谈到学习和教导的主要性时,所要表白的意思。我们需要跳出“事件应该如何;这样的理论探讨,并且采用举动,以确保他们有机会介入进来。

我们站在一个漂亮新世界的进口。这是一个令人高兴的、同时充斥了不确定性的世界,而你们是先行者。我祝愿你们。

问答环节

中国科技大咖、科学家、投资家和网友提问霍金(在百万网友的关注中筛选出的七个问题)

Q1:翻新工场的CEO李开复:

互联网巨头领有巨量的数据,而这些数据会给他们各种以用户隐私和利益换取暴利的机会。在巨大的利益引诱下,他们是无法自律的。而且,这种行动也会导致小公司和创业者更难立异。您常谈到如何约束人工智能,但更难的是如何约束人本身。您认为我们应该如何束缚这些巨头?

A1:

据我懂得,许多公司仅将这些数据用于统计分析,但任何波及到私家信息的使用都应该被禁止。会有助于隐私维护的是,如果互联网上所有的信息,均通过基于量子技术加密,这样互联网公司在一定时间内便无法破解。但安全服务会反对这个做法。

Q2:猎豹移动CEO傅盛:

“灵魂会不会是量子的一种存在状态?或者是高维空间里的另一个表示?

A2:

近来人工智能的发展,比如电脑在国际象棋和围棋的比赛中克服人脑,都显示出人脑和电脑并没有本质差异。这点上我和我的同事罗杰·彭罗斯正好相反。会有人认为电脑有灵魂吗?对我而言,灵魂这个说法是一个基督教的概念,它和下世接洽在一起。我认为这是一个童话故事。

Q3:百度总裁张亚勤:

“人类视察和形象世界的方法不断演进,从早期的察看和估算,到牛顿定律和爱因斯坦方程式, 到今天数据驱动的计算和人工智能,下一个是什么?;

A3:

我们需要一个新的量子理论,将重力和其他天然界的其它力气整合在一起。许多人宣称这是弦理论,但我对此表现猜忌,目前独一的揣测是,时空有十个维度。

Q4:斯坦福大学物理学教学张首晟:

“假如让你告知外星人我们人类取得的最高成绩,写在一张明信片的反面,您会写什么?;

A4:

告诉外星人关于美,或者任何可能代表最高艺术造诣的艺术情势都是无益的。因为这是人类特有的。我会告诉他们哥德尔不齐备定理和费马大定理。这才是外星人可以理解的事情。

Q5:

“我们盼望提倡科学精力,贯串GMIC全球九站,清您推举三本书,让科技届的友人们更好的理解科学及科学的未来。;

A5:

他们应当去写书而不是读书。只有当一个人对于某件事能写出一本书,才代表他完整懂得了这件事。

Q6:微博用户:

“您认为一个人毕生中最应该做的一件事和最不应当做的一件事是什么?;

A6:

我们毫不应当废弃,我们都应当尽可能多地去理解这个世界。

Q7:微博用户:

“人类在漫漫的历史长河中,反复着一次又一次的革命与活动。从石器、蒸汽、电气……你认为下一次的革命会是由什么驱动的?;

A7:

(我认为是)计算机科学的发展,包含人工智能和量子计算。科技已经成为我们生活中重要的一部分,但未来几十年里,它会逐步浸透到社会的每一个方面,为我们提供智能地支持和倡议,在医疗、工作、教育和科技等众多领域。但是我们必需要确保是我们来掌控人工智能,而非它(掌控)我们。

Q8:音乐人、投资者胡海泉:

“如果星际移民技术的成熟窗口期迟到,有没有完全解决不了的内发灾害导致人类灭绝?抛开陨星撞地球这样的外来灾害。;

A8:

是会有这个内发劫难的。人口多余、疾病、战斗、饥馑、气象变化和水资源匮乏, 人类有能力解决这些危机。但很惋惜,这些危机还重大威胁着我们在地球上的生存,这些危机都是可以解决的,但目前还没有。

英文演讲全文:

Over my lifetime, I have seen very significant societal changes. Probably one of the most significant, and one that is increasingly concerning people today, is the rise of artificial intelligence.

In short, I believe that the rise of powerful AI, will be either the best thing, or the worst, ever to happen to humanity.  

I have to say now, that we do not yet know which. But we should do all we can, to ensure that its future development benefits us, and our environment. We have no other option. I see the development of AI, as a trend with its own problems that we know must be dealt with, now and into the future.

The progress in AI research and development is swift. And perhaps we should all stop for a moment, and focus our research, not only on making AI more capable, but on maximizing its societal benefit.

Such considerations motivated the American Association for Artificial Intelligence's, two thousand and eight to two thousand and nine, Presidential Panel on Long-Term AI Futures, which up to recently had focused largely on techniques, that are neutral with respect to purpose. 

But our AI systems must do what we want them to do. Inter-disciplinary research can be a way forward: ranging from economics, law, and philosophy, to computer security, formal methods, and of course various branches of AI itself.

  

Everything that civilization has to offer, is a product of human intelligence, and I believe there is no real difference between what can be achieved by a biological brain, and what can be achieved by a computer. 

It therefore follows that computers can, in theory, emulate human intelligence, and exceed it. But we don’t know. So we cannot know if we will be infinitely helped by AI, or ignored by it and side-lined, or conceivably destroyed by it. 

Indeed, we have concerns that clever machines will be capable of undertaking work currently done by humans, and swiftly destroy millions of jobs.

While primitive forms of artificial intelligence developed so far, have proved very useful, I fear the consequences of creating something that can match or surpass humans. AI would take off on its own, and re-design itself at an ever increasing rate. Humans, who are limited by slow biological evolution, couldn't compete, and would be superseded. It will bring great disruption to our economy.

And in the future, AI could develop a will of its own, a will that is in conflict with ours. Although I am well-known as an optimist regarding the human race, others believe that humans can command the rate of technology for a decently long time, and that the potential of AI to solve many of the world's problems will be realised. I am not so sure.

In January 2015, I, along with the technological entrepreneur, Elon Musk, and many other AI experts, signed an open letter on artificial intelligence, calling for serious research on its impact on society. 

In the past, Elon Musk has warned that super human artificial intelligence, is possible of providing incalculable benefits, but if deployed incautiously, will have an adverse effect on the human race. 

He and I, sit on the scientific advisory board for the Future of Life Institute, an organization working to mitigate existential risks facing humanity, and which drafted the open letter. This called for concrete research on how we could prevent potential problems, while also reaping the potential benefits AI offers us, and is designed to get AI researchers and developers to pay more attention to AI safety.

In addition, for policymakers and the general public, the letter is meant to be informative, but not alarmist. We think it is very important, that everybody knows that AI researchers are seriously thinking about these concerns and ethical issues. 

For example, AI has the potential to eradicate disease and poverty, but researchers must work to create AI that can be controlled. The four-paragraph letter, titled Research Priorities for Robust and Beneficial Artificial Intelligence, an Open Letter, lays out detailed research priorities in the accompanying twelve-page document.

For the last 20 years or so, AI has been focused on the problems surrounding the construction of intelligent agents, systems that perceive and act in some environment. In this context, intelligence is related to statistical and economic notions of rationality. Colloquially, the ability to make good decisions, plans, or inferences.

As a result of this recent work, there has been a large degree of integration and cross-fertilisation among AI, machine learning, statistics, control theory, neuroscience, and other fields. The establishment of shared theoretical frameworks, combined with the availability of data and processing power, has yielded remarkable successes in various component tasks, such as speech recognition, image classification, autonomous vehicles, machine translation, legged locomotion, and question-answering systems.

As development in these areas and others, moves from laboratory research to economically valuable technologies, a virtuous cycle evolves, whereby even small improvements in performance, are worth large sums of money, prompting further and greater investments in research. 

There is now a broad consensus that AI research is progressing steadily, and that its impact on society is likely to increase. The potential benefits are huge, since everything that civilization has to offer, is a product of human intelligence; we cannot predict what we might achieve, when this intelligence is magnified by the tools AI may provide. 

But, and as I have said, the eradication of disease and poverty is not unfathomable. Because of the great potential of AI, it is important to research how to reap its benefits, while avoiding potential pitfalls.

Artificial intelligence research is now progressing rapidly. And this research can be discussed as short-term and long-term. Some short-term concerns relate to autonomous vehicles, from civilian drones and self-driving cars. For example, a self-driving car may, in an emergency, have to decide between a small risk of a major accident, and a large probability of a small accident. 

Other concerns relate to lethal intelligent autonomous weapons. Should they be banned. If so, how should autonomy be precisely defined. If not, how should culpability for any misuse or malfunction be apportioned. Other issues include privacy concerns, as AI becomes increasingly able to interpret large surveillance datasets, and how to best manage the economic impact of jobs displaced by AI.

Long-term concerns, comprise primarily of the potential loss of control of AI systems, via the rise of super-intelligences that do not act in accordance with human wishes, and that such powerful systems would threaten humanity. Are such dystopic outcomes possible. 

If so, how might these situations arise. What kind of investments in research should be made, to better understand and to address the possibility of the rise of a dangerous super-intelligence, or the occurrence of an intelligence explosion.

Existing tools for harnessing AI, such as reinforcement learning, and simple utility functions, are inadequate to solve this. Therefore more research is necessary to find and validate a robust solution to the control problem.

Recent landmarks, such as the self-driving cars already mentioned, or a computer winning at the game of Go, are signs of what is to come.  Enormous levels of investment are pouring into this technology. 

The achievements we have seen so far, will surely pale against what the coming decades will bring, and we cannot predict what we might achieve, when our own minds are amplified by AI.  

Perhaps with the tools of this new technological revolution, we will be able to undo some of the damage done to the natural world by the last one, industrialisation. Every aspect of our lyves will be transformed. In short, success in creating AI, could be the biggest event in the history of our civilisation.

But it could also be the last, unless we learn how to avoid the risks.  I have said in the past that the development of full AI, could spell the end of the human race, such as the ultimate use of powerful autonomous weapons. Earlier this year, I, along with other international scientists, supported the United Nations convention to negotiate a ban on nuclear weapons. 

We await the outcome with nervous anticipation. Currently, nine nuclear powers have access to roughly 14,000 nuclear weapons, any one of which can obliterate cities, contaminate wide swathes of land with radioactive fall-out, and the most horrible hazard of all, cause a nuclear-induced winter, in which the fires and smoke might trigger a global mini-ice age. 

The result is a complete collapse of the global food system, and apocalyptic unrest, potentially killing most people on earth. We scientists bear a special responsibility for nuclear weapons, since it was scientists who invented them, and discovered that their effects are even more horrific than first thought.

At this stage, I may have possibly frightened you all here today, with talk of doom. I apologise. But it is important that you, as attendees to today's conference, recognise the position you hold in influencing future research and development of today's technology. 

I believe that we join together, to call for support of international treaties, or signing letters presented to individual governmental powers. Technology leaders and scientists are doing what they can, to obviate the rise of uncontrollable AI.

In October last year, I opened a new center in Cambridge, England, which will attempt to tackle some of the open-ended questions raised by the rapid pace of development in AI research. The Leverhulme Centre for the Future of Intelligence, is a multi-disciplinary institute, dedicated to researching the future of intelligence, as crucial to the future of our civilisation and our species. We spend a great deal of time studying history, which let's face it, is mostly the history of stupidity. 

So it's a welcome change, that people are studying instead the future of intelligence. We are aware of the potential dangers, but I am at heart an optimist, and believe that the potential benefits of creating intelligence are huge. Perhaps with the tools of this new technological revolution, we will be able to undo some of the damage done to the natural world, by industrialisation.

Every aspect of our lives will be transformed. My colleague at the institute, Huw Price, has acknowledged that the center came about partially as a result of the university’s Centre for Existential Risk. That institute examines a wider range of potential problems for humanity, while the Leverhulme Centre has a more narrow focus.

Recent developments in the advancement of AI, include a call by the European Parliament for drafting a set of regulations, to govern the yooss and creation of robots and AI. Somewhat surprisingly, this includes a form of electronic personhood, to ensure the rights and responsibilities for the most capable and advanced AI. 

A European Parliament spokesman has commented, that as a growing number of areas in our daily lyves are increasingly affected by robots, we need to ensure that robots are, and will remain, in the service of humans.

The report as presented to MEPs, makes it clear that it believes the world is on the cusp of a new industrial robot revolution. It examines whether or not providing legal rights for robots as electronic persons, on a par with the legal definition of corporate personhood, would be permissible. 

But stresses that at all times, researchers and designers should ensure all robotic design incorporates a kill switch. This didn't help the scientists on board the spaceship with Hal, the malfunctioning robotic computer in Kubrick’s two thousand and one, a Space Odyssey, but that was fiction. We deal with fact. Lorna Brazell, a partner at the multinational law firm Osborne Clarke, says in the report, that we don’t give whales and gorillas personhood, so there is no need to jump at robotic personhood. 

But the wariness is there. The report acknowledges the possibility that within the space of a few decades, AI could surpass human intellectual capacity, and challenge the human robot relationship. Finally, the report calls for the creation of a European agency for robotics and AI, that can provide technical, ethical, and regulatory expertise. If MEPs vote in favor of legislation, the report will go to the European Commission, which has three months to decide what legislative steps it will take. 

We too, have a role to play in making sure the next generation has not just the opportunity, but the determination, to engage fully with the study of science at an early level, so that they can go on to fulfil their potential, and create a better world for the whole human race. 

This is what I meant, when I was talking to you just now about the importance of learning and education. We need to take this beyond a theoretical discussion of how things should be, and take action, to make sure they have the opportunity to get on board. We stand on the threshold of a brave new world. It is an exciting, if precarious place to be, and you are the pioneers. I wish you well.

 

 

Professor Hawking, we have learned so much from your insight.

Next I’m going to ask some questions. These are from Chinese scientists and entrepreneurs.

Kai-Fu Lee, CEO of Sinovation Ventures:

"The large internet companies have access to massive databases, which allows them to make huge strides in AI by violating user's privacy. These companies can’t truly discipline themselves as they are lured by huge economic interests. This vastly disproportionate access to data could cause small companies and startups to fail to innovate. You have mentioned numerous times that we should restrain artificial intelligence, but it’s much harder to restrain humans. What do you think we can do to restrain the large internet companies?"

As I understand it,the companies are using the data only for statistical purposes,but use of any personal information should be banned. It would help privacy, if all material on the internet, were encrypted by quantum cryptography with a code, that the internet companies could not break in a reasonable time. But the security services would object.

Professor, the second question is from Fu Sheng, CEO, Cheetah Mobile:

“Does the human soul exist as a form of quantum or another form of higher dimensional space?"

I believe that recent advances in AI, such as computers winning at chess and Go, show that there is no essential difference between the human brain and a computer. Contrary to the opinion of my colleague Roger Penrose. Would one say a computer has a soul. In my opinion, the notion of an individual human soul is a Christian concept, linked to the afterlife which I consider to be a fairy story.

Professor,the third question is from Ya-Qin Zhang, President, Baidu:  

“The way human beings observe and abstract the universe is constantly evolving, from observation and estimation to Newton's law and Einstein’s equation , and now data-driven computation and AI . What is next;

We need a new quantum theory, which unifies gravity with the other forces of nature. Many people claim that it is string theory, but I have my doubts. So far about the only prediction is that space-time has ten dimensions.

Professor, the forth question is from Zhang Shoucheng , Professor of Physics, Stanford University:

“If you were to tell aliens about the highest achievements of our human civilization on the back of one envelope, what would you write ?;

It is no good telling aliens about beauty or any other possible art form that we might consider to be the highest artistic achievement,because these are very human specific. Instead I would write about Godel’s Incompleteness Theorems and Fermat’s Last Theorem. These are things aliens would understand

The next question is from myself:

“We wish to promote the scientific spirit at all 9 GMIC conferences globally. What three books do you recommend technology leaders read to better understand the coming future and the science that is driving it?;

They should be writing books not reading them. One fully understands something only when one has written a book about it.

  

The next question is from Weibo user:

What is the one thing we should never do in life, and the one thing we should all do?

We should never give up, and we should all strive to understand as much as we can.

The next question is also from Weibo user:

“Human beings have experienced many evolutions ,for example, the Stone Age, the age of steam to the age of electricity. What do you think will drive the next evolution?;

Advances in computer science, including artificial intelligence and quantum computing. Technology already forms a major part of our lives but in the coming decades, it will permeate every aspect of our society .intelligently supporting and advising us in many areas , including healthcare work education and science. But we must make sure we control AI not it us.

Professor Hawking,the last question is from Hai Quan , Musician and VC:

“If the technology is not mature yet for interstellar immigrants, do human beings have unsolvable challenges that could lead to human extinction apart from external catastrophes like asteroid hitting earth?;

Yes. over-population, disease, war, famine, climate change and lack of water. It is within the power of man to solve these crises, but unfortunately these remain serious threats to our continued present on earth. These are all solvable, but so far have not been.

·END·

 


相关的主题文章: