0% found this document useful (0 votes)
1K views17 pages

English Sba

The document provides a history of artificial intelligence from its early conceptualization in science fiction to modern developments. It discusses key events like Alan Turing proposing the possibility of machine intelligence, the 1956 Dartmouth conference that coined the term "artificial intelligence", successes in the 1950s-1970s with programs like the Logic Theorist, and later setbacks due to lack of computational power. Funding increased again in the 1980s with the development of new algorithms and expert systems, though goals were still not fully achieved. Landmark achievements continued in the 1990s-2000s even without major public/government attention.

Uploaded by

Hope Grant
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
1K views17 pages

English Sba

The document provides a history of artificial intelligence from its early conceptualization in science fiction to modern developments. It discusses key events like Alan Turing proposing the possibility of machine intelligence, the 1956 Dartmouth conference that coined the term "artificial intelligence", successes in the 1950s-1970s with programs like the Logic Theorist, and later setbacks due to lack of computational power. Funding increased again in the 1980s with the development of new algorithms and expert systems, though goals were still not fully achieved. Landmark achievements continued in the 1990s-2000s even without major public/government attention.

Uploaded by

Hope Grant
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 17

ENGLISH LANGUAGE SCHOOL BASED ASSESSMENT ON SCAMMING

Candidate name: Hope Grant

School: Holland High School

Teacher: Edith Daley

Centre number: 100366

Candidate number:

Territory: Jamaica

Year: 2023
Table of Contents

Theme and topic ……………………………………………………………………...3

Plan of Investigation …………………………………………………………………4

Participation Measure 2nd draft …………………….………………………………...5

Written Report ……………………………………………………………………….6

ARTIFACTS………………………………………………………………………..8

Newspaper Article ......................................................................................................8

Picture ........................................................................................................................16

Youtube Video ...........................................................................................................17

REFLECTIONS …………………………………………………………………….19

Reflection 1 ………………………………………………………………………....20

Reflection 2 …………………………………………………………………………21

Reflection 3 …………………………………………………………………………22

Plan for Oral Presentation…………………………………………..……………….24

Work Cited…………………………………………………………………………..25

APPENDICES……………………………………………………………………….26

Appendix # 1-Oral Presentation ….…………...…………………………………….27

Appendix # 2-Participation Measure 1st draft ………………………………………30


Theme and Topic

Theme: Scamming

Topic: What Influences people into Scamming in the World

PLAN OF INVESTIGATION
The theme that I have chosen to do my research on is “Scamming”. The topic that I will be

doing is “What Influences people into Scamming in the World”. I became interested in this

topic after noticing the negative effects of Scamming among the young people in my country.

My expected benefit of this topic is to help widen my vocabulary and extent my thinking

skills. I intend to collect information on this topic by browsing search engines like

chorme,google to research a newspaper Article a song and video on the issues of Scamming .

The skill that I will be using, to write my reflections, written report and oral presentation, I

will use is summary skills, public speaking and critical thinking skills.

PARTICIPATION MEASURE – 2nd DRAFT

Rating Criteria Preliminary Student’s Records Of Final Rating

Rating Teacher’s

Comment/suggestions
1- Agree

1-  Agree 0- Uncertain

0-  Uncertain -1- Disagree

-1- Disagree
I know what I am expected          1

to do to work well in a

team.

It is easy to focus on what is          1

taking place in my group

In the group, I feel self-           1

confidence, self-esteem and

efficacy.

I know how to manage how           1

I behave in the group.

I always manage my tasks           1

on time and thoroughly.

Total: 5/5 Total: 5/5

WRITTEN REPORT
ARTIFACTS
ARTIFACT #1- NEWSPAPER ARTICLE

AUGUST 28, 2017

BLOG, SPECIAL EDITION ON ARTIFICIAL INTELLIGENCE


The History of Artificial Intelligence

By Rockwell Anyoha

Can Machines Think?

In the first half of the 20th century, science fiction familiarized the world with the concept of

artificially intelligent robots. It began with the “heartless” Tin man from the Wizard of Oz and

continued with the humanoid robot that impersonated Maria in Metropolis. By the 1950s, we

had a generation of scientists, mathematicians, and philosophers with the concept of artificial

intelligence (or AI) culturally assimilated in their minds. One such person was Alan Turing, a

young British polymath who explored the mathematical possibility of artificial intelligence.

Turing suggested that humans use available information as well as reason in order to solve

problems and make decisions, so why can’t machines do the same thing? This was the logical

framework of his 1950 paper, Computing Machinery and Intelligence in which he discussed

how to build intelligent machines and how to test their intelligence.

Making the Pursuit Possible

Unfortunately, talk is cheap. What stopped Turing from getting to work right then and there?

First, computers needed to fundamentally change. Before 1949 computers lacked a key

prerequisite for intelligence: they couldn’t store commands, only execute them. In other

words, computers could be told what to do but couldn’t remember what they did. Second,

computing was extremely expensive. In the early 1950s, the cost of leasing a computer ran up

to $200,000 a month. Only prestigious universities and big technology companies could

afford to dillydally in these uncharted waters. A proof of concept as well as advocacy from
high profile people were needed to persuade funding sources that machine intelligence was

worth pursuing.

The Conference that Started it All

Five years later, the proof of concept was initialized through Allen Newell, Cliff Shaw, and

Herbert Simon’s, Logic Theorist. The Logic Theorist was a program designed to mimic the

problem solving skills of a human and was funded by Research and Development (RAND)

Corporation. It’s considered by many to be the first artificial intelligence program and was

presented at the Dartmouth Summer Research Project on Artificial Intelligence (DSRPAI)

hosted by John McCarthy and Marvin Minsky in 1956. In this historic conference, McCarthy,

imagining a great collaborative effort, brought together top researchers from various fields

for an open ended discussion on artificial intelligence, the term which he coined at the very

event. Sadly, the conference fell short of McCarthy’s expectations; people came and went as

they pleased, and there was failure to agree on standard methods for the field. Despite this,

everyone whole-heartedly aligned with the sentiment that AI was achievable. The

significance of this event cannot be undermined as it catalysed the next twenty years of AI

research.

Roller Coaster of Success and Setbacks

From 1957 to 1974, AI flourished. Computers could store more information and became

faster, cheaper, and more accessible. Machine learning algorithms also improved and people

got better at knowing which algorithm to apply to their problem. Early demonstrations such

as Newell and Simon’s General Problem Solver and Joseph Weizenbaum’s ELIZA showed

promise toward the goals of problem solving and the interpretation of spoken language
respectively. These successes, as well as the advocacy of leading researchers (namely the

attendees of the DSRPAI) convinced government agencies such as the Defense Advanced

Research Projects Agency (DARPA) to fund AI research at several institutions. The

government was particularly interested in a machine that could transcribe and translate

spoken language as well as high throughput data processing. Optimism was high and

expectations were even higher. In 1970 Marvin Minsky told Life Magazine, “From three to

eight years we will have a machine with the general intelligence of an average human being.”

However, while the basic proof of principle was there, there was still a long way to go before

the end goals of natural language processing, abstract thinking, and self-recognition could be

achieved.

Breaching the initial fog of AI revealed a mountain of obstacles. The biggest was the lack of

computational power to do anything substantial: computers simply couldn’t store enough

information or process it fast enough. In order to communicate, for example, one needs to

know the meanings of many words and understand them in many combinations. Hans
Moravec, a doctoral student of McCarthy at the time, stated that “computers were still

millions of times too weak to exhibit intelligence.”  As patience dwindled so did the funding,

and research came to a slow roll for ten years.

In the 1980’s, AI was reignited by two sources: an expansion of the algorithmic toolkit, and a

boost of funds. John Hopfield and David Rumelhart popularized “deep learning” techniques

which allowed computers to learn using experience. On the other hand Edward Feigenbaum

introduced expert systems which mimicked the decision making process of a human expert.

The program would ask an expert in a field how to respond in a given situation, and once this

was learned for virtually every situation, non-experts could receive advice from that program.

Expert systems were widely used in industries. The Japanese government heavily funded

expert systems and other AI related endeavours as part of their Fifth Generation Computer

Project (FGCP). From 1982-1990, they invested $400 million dollars with the goals of

revolutionizing computer processing, implementing logic programming, and improving

artificial intelligence. Unfortunately, most of the ambitious goals were not met. However, it

could be argued that the indirect effects of the FGCP inspired a talented young generation of

engineers and scientists. Regardless, funding of the FGCP ceased, and AI fell out of the

limelight.

Ironically, in the absence of government funding and public hype, AI thrived. During the

1990s and 2000s, many of the landmark goals of artificial intelligence had been achieved. In

1997, reigning world chess champion and grand master Gary Kasparov was defeated by

IBM’s Deep Blue, a chess playing computer program. This highly publicized match was the

first time a reigning world chess champion loss to a computer and served as a huge step

towards an artificially intelligent decision making program. In the same year, speech

recognition software, developed by Dragon Systems, was implemented on Windows. This


was another great step forward but in the direction of the spoken language interpretation

endeavour. It seemed that there wasn’t a problem machines couldn’t handle. Even human

emotion was fair game as evidenced by Kismet, a robot developed by Cynthia Breazeal that

could recognize and display emotions.

Time Heals All Wounds

We haven’t gotten any smarter about how we are coding artificial intelligence, so what

changed? It turns out, the fundamental limit of computer storage that was holding us back 30

years ago was no longer a problem. Moore’s Law, which estimates that the memory and

speed of computers doubles every year, had finally caught up and in many cases, surpassed

our needs. This is precisely how Deep Blue was able to defeat Gary Kasparov in 1997, and

how Google’s Alpha Go was able to defeat Chinese Go champion, Ke Jie, only a few months

ago. It offers a bit of an explanation to the roller coaster of AI research; we saturate the

capabilities of AI to the level of our current computational power (computer storage and

processing speed), and then wait for Moore’s Law to catch up again.

Artificial Intelligence is Everywhere

We now live in the age of “big data,” an age in which we have the capacity to collect huge

sums of information too cumbersome for a person to process. The application of artificial

intelligence in this regard has already been quite fruitful in several industries such as

technology, banking, marketing, and entertainment. We’ve seen that even if algorithms don’t

improve much, big data and massive computing simply allow artificial intelligence to learn

through brute force. There may be evidence that Moore’s law is slowing down a tad, but the
increase in data certainly hasn’t lost any momentum. Breakthroughs in computer science,

mathematics, or neuroscience all serve as potential outs through the ceiling of Moore’s Law.

The Future

So what is in store for the future? In the immediate future, AI language is looking like the

next big thing. In fact, it’s already underway. I can’t remember the last time I called a

company and directly spoke with a human. These days, machines are even calling me! One

could imagine interacting with an expert system in a fluid conversation, or having a

conversation in two different languages being translated in real time. We can also expect to

see driverless cars on the road in the next twenty years (and that is conservative). In the long

term, the goal is general intelligence, that is a machine that surpasses human cognitive

abilities in all tasks. This is along the lines of the sentient robot we are used to seeing in

movies. To me, it seems inconceivable that this would be accomplished in the next 50 years.

Even if the capability is there, the ethical questions would serve as a strong barrier against

fruition. When that time comes (but better even before the time comes), we will need to have

a serious conversation about machine policy and ethics (ironically both fundamentally human

subjects), but for now, we’ll allow AI to steadily improve and run amok in society.

Rockwell Anyoha is a graduate student in the department of molecular biology with a

background in physics and genetics. His current project employs the use of machine learning

to model animal behavior. In his free time, Rockwell enjoys playing soccer and debating

mundane topics.

This article is part of a Special Edition on Artificial Intelligence.


For more information:

Brief Timeline of AI

https://www.livescience.com/47544-history-of-a-i-artificial-intelligence-infographic.html

Complete Historical Overview

http://courses.cs.washington.edu/courses/csep590/06au/projects/history-ai.pdf

Dartmouth Summer Research Project on Artificial Intelligence

https://www.aaai.org/ojs/index.php/aimagazine/article/view/1904/1802

Future of AI

https://www.technologyreview.com/s/602830/the-future-of-artificial-intelligence-and-

cybernetics/

Discussion on Future Ethical Challenges Facing AI

http://www.bbc.com/future/story/20170307-the-ethical-challenge-facing-artificial-

intelligence

This work by SITNBoston is licensed under a Creative Commons Attribution-

NonCommercial-ShareAlike 4.0 International License.


 Unless otherwise indicated, attribute to the author or graphics designer and SITNBoston,

linking back to this page if possible.

Proudly powered by WordPress  Theme: Canard by Automattic.

ARTIFACT #2- PICTURE

Artificial intelligence In Everyday Life


ARTIFACT #3- VIDEO

Video on the latest AI and Deep learning Technology Innovations

You might also like