Skip to main content

ARTIFICIAL INTELLIGENCE

                           ARTIFICIAL INTELLIGENCE



In computer science, artificial intelligence (AI), sometimes called machine intelligence, is intelligence demonstrated by machines, in contrast to the natural intelligence displayed by humans and animals. Leading AI textbooks define the field as the study of "intelligent agents": any device that perceives its environment and takes actions that maximize its chance of successfully achieving its goals. Colloquially, the term "artificial intelligence" is often used to describe machines (or computers) that mimic "cognitive" functions that humans associate with the human mind, such as "learning" and "problem solving".
As machines become increasingly capable, tasks considered to require "intelligence" are often removed from the definition of AI, a phenomenon known as the AI effect. A quip in Tesler's Theorem says "AI is whatever hasn't been done yet." For instance, optical character recognition is frequently excluded from things considered to be AI , having become a routine technology. Modern machine capabilities generally classified as AI include successfully understanding human speech, competing at the highest level in  strategic game systems (such as chess and Go), autonomously operating cars, intelligent routing in content delivery networks, and military simulations.
Artificial intelligence was founded as an academic discipline in 1955, and in the years since has experienced several waves of optimism, followed by disappointment and the loss of funding (known as an "AI winter"), followed by new approaches, success and renewed funding. For most of its history, AI research has been divided into subfields that often fail to communicate with each other. These sub-fields are based on technical considerations, such as particular goals (e.g. "robotics" or "machine learning"), the use of particular tools ("logic" or artificial neural networks), or deep philosophical differences. Subfields have also been based on social factors (particular institutions or the work of particular researchers).
The major limitation in defining AI as simply "building machines that are intelligent" is that it doesn't actually explain what artificial intelligence is? What makes a machine intelligent?
In their groundbreaking textbook Artificial Intelligence: A Modern Approach, authors Stuart Russell and Peter Norvig approach the question by unifying their work around the theme of intelligent agents in machines. With this in mind, AI is "the study of agents that receive percepts from the environment and perform actions." (Russel and Norvig viii)
Norvig and Russell go on to explore four different approaches that have historically defined the field of AI: 
1.     Thinking humanly
2.   Thinking rationally
3.   Acting humanly 
4.   Acting rationally
The first two ideas concern thought processes and reasoning, while the others deal with behavior. Norvig and Russell focus particularly on rational agents that act to achieve the best outcome, noting "all the skills needed for the Turing Test also allow an agent to act rationally." (Russel and Norvig 4).
Patrick Winston, the Ford professor of artificial intelligence and computer science at MIT, defines AI as  "algorithms enabled by constraints, exposed by representations that support models targeted at loops that tie thinking, perception and action together."
In the near term, the goal of keeping AI’s impact on society beneficial motivates research in many areas, from economics and law to technical topics such as verification, validity, security and control. Whereas it may be little more than a minor nuisance if your laptop crashes or gets hacked, it becomes all the more important that an AI system does what you want it to do if it controls your car, your airplane, your pacemaker, your automated trading system or your power grid. Another short-term challenge is preventing a devastating arms race in lethal autonomous weapons.
In the long term, an important question is what will happen if the quest for strong AI succeeds and an AI system becomes better than humans at all cognitive tasks. As pointed out by I.J. Good in 1965, designing smarter AI systems is itself a cognitive task. Such a system could potentially undergo recursive self-improvement, triggering an intelligence explosion leaving human intellect far behind. By inventing revolutionary new technologies, such a superintelligence might help us eradicate war, disease, and poverty, and so the creation of strong AI might be the biggest event in human history. Some experts have expressed concern, though, that it might also be the last, unless we learn to align the goals of the AI with ours before it becomes superintelligent.
There are some who question whether strong AI will ever be achieved, and others who insist that the creation of superintelligent AI is guaranteed to be beneficial. At FLI we recognize both of these possibilities, but also recognize the potential for an artificial intelligence system to intentionally or unintentionally cause great harm. We believe research today will help us better prepare for and prevent such potentially negative consequences in the future, thus enjoying the benefits of AI while avoiding pitfalls.
HOW CAN AI BE DANGEROUS?
Most researchers agree that a superintelligent AI is unlikely to exhibit human emotions like love or hate, and that there is no reason to expect AI to become intentionally benevolent or malevolent. Instead, when considering how AI might become a risk, experts think two scenarios most likely:
1. The AI is programmed to do something devastating: Autonomous weapons are artificial intelligence systems that are programmed to kill. In the hands of the wrong person, these weapons could easily cause mass casualties. Moreover, an AI arms race could inadvertently lead to an AI war that also results in mass casualties. To avoid being thwarted by the enemy, these weapons would be designed to be extremely difficult to simply “turn off,” so humans could plausibly lose control of such a situation. This risk is one that’s present even with narrow AI, but grows as levels of AI intelligence and autonomy increase.
2.                        The AI is programmed to do something beneficial, but it develops a destructive method for achieving its goal: This can happen whenever we fail to fully align the AI’s goals with ours, which is strikingly difficult. If you ask an obedient intelligent car to take you to the airport as fast as possible, it might get you there chased by helicopters and covered in vomit, doing not what you wanted but literally what you asked for. If a superintelligent system is tasked with a ambitious geoengineering project, it might wreak havoc with our ecosystem as a side effect, and view human attempts to stop it as a threat to be met.
As these examples illustrate, the concern about advanced AI isn’t malevolence but competence. A super-intelligent AI will be extremely good at accomplishing its goals, and if those goals aren’t aligned with ours, we have a problem. You’re probably not an evil ant-hater who steps on ants out of malice, but if you’re in charge of a hydroelectric green energy project and there’s an anthill in the region to be flooded, too bad for the ants. A key goal of AI safety research is to never place humanity in the position of those ants.



Comments

Popular posts from this blog

Ammonium phosphate

                                            Ammonium phosphate Ammonium phosphate  is an ammonium  salt of orthophosphoric acid. It is a highly unstable compound with the  formula (NH 4 ) 3 PO 4 . Because of its instability, it is elusive and of no commercial value. A related "double salt", (NH 4 ) 3 PO 4 . (NH 4 ) 2 HPO 4  is also recognized but is too unstable for practical use. Both triammonium salts evolve ammonia. In contrast to the unstable nature of the triammonium salts, the  diammonium phosphate (NH 4 ) 2 HPO 4  monoammonium salt (NH 4 )H 2 PO 4,  are stable materials that are commonly used as fertilizers to provide plants with fixed nitrogen and phosphorus. PREPARATION OH AMMONIUM TRIPHOSPHATE Triammonium phosphate can be prepared in the laboratory by treating 85% phosphoric acid with 30% ammonia solution: ...

class 9 chapter 1 ncert notes

Class-9 Chapter-1 Matters in our surroundings. 1) What is matter? Ans.   Matter is anything that occupied space that have mass and volume and offer residence is called matter. OR...... Matter is anything that occupies mass and volume. OR..... We can say that anything that has occupies space and mass of can be felt by our one or more sense organ is called matter. NOTE- 1. The SI unit of mass is kilogram (kg). 2. The SI unit of volume is cubic meter (m 3 ). 3. The common unit of measuring volume is litre (L).        (1L=1dm 3 , 1L=1000mL, 1mL= 1cm 3 ) For example- Full form: 1. MKS- Metre Kilogram Second 2. CGS- Centi Gram Second 3. FPS- Foot Pound Second Classification of matter- 1. Solid 2. Liquid 3. Gas       Solid- solid has definite shape and volume. Liquid- Liquid has definite volume but not definite shape. Gas- Gas has neither definite shape nor definite vo...

Akhilesh yadav

                                                    Akhilesh Yadav, son of Mulayam Singh Yadav is a dynamic young leader of Samajwadi leader. In March 2012, at the age of 38 years, he became the youngest person to hold the post. Akhilesh Yadav is now the president of National President of Samajwadi Party. Before the elections there was tussle in the Yadav family while Shivpal Singh wanted to push his candidates, Akhilesh was adamant on finalizing the candidate list. After much bickering in the family he assumed the command of the Samajwadi Party by dislodging his father Mulayam Singh Yadav and sidelining his uncle Shivpal Singh Yadav. He went on to forge an alliance with the congress after securing the command of the party before the assembly elections of Uttar Pradesh and campaigned for the party. He harped upon the development and asked peo...