0% found this document useful (0 votes)
7 views12 pages

RTIT

The document outlines the examination structure for BCA III (Semester V) at Solapur University, focusing on Recent Trends in Information Technology and Data Warehouse and Mining. It includes multiple-choice questions, short answer questions, and essay prompts covering various IT topics such as virtualization, big data, Hadoop architecture, and machine learning. The exam is designed to assess students' understanding and application of contemporary IT concepts.

Uploaded by

anitarooge
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views12 pages

RTIT

The document outlines the examination structure for BCA III (Semester V) at Solapur University, focusing on Recent Trends in Information Technology and Data Warehouse and Mining. It includes multiple-choice questions, short answer questions, and essay prompts covering various IT topics such as virtualization, big data, Hadoop architecture, and machine learning. The exam is designed to assess students' understanding and application of contemporary IT concepts.

Uploaded by

anitarooge
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd

SOLAPUR UNIVERSITY, SOLAPUR

BCA III (Semester V)(CBCS) Examination April 2020


Recent Trends in Information Technology
Day & Date:
Time Duration: 2Hr Total Marks: 70
Instructions :
1) All Questions are compulsory.
2) Figures to right indicate full marks.
Q. 1) A ) Choose correct Alternatives (14)
1) Which of the following is Type 2 VM?
a) VirtualLogix VLX b) VMware ESX c) Xen d) LynxSecure
2) What are the 3 R's of the environment?
a) Reduce, remove and recycle b) Reduce, rewash and retain
c) Reduce, reuse and recycle d) None of these
3) Which of the following type of virtualization is also characteristic of cloud computing?
a) Storage b) Application
c) CPU d) All of the mentioned
4) When a client communicates with the HDFS file system, it needs to communicate with
a) only the namenode b) only the data node
c) both the namenode and datanode d)None of these
5) In which category of SAAS service CRM(Customer Relationship Management) software fall
a) Customer Services b) Communication Services
c) Infrastructure Services d) Business Services
6) The sample is subset of _______
a) Data b) Group c) Population d) Distribution
7)What is Machine learning?
a) The autonomous acquisition of knowledge through the use of computer programs
b) The autonomous acquisition of knowledge through the use of manual programs
c) The selective acquisition of knowledge through the use of computer programs
d) The selective acquisition of knowledge through the use of manual programs
8) The population consists of the results of repeated trials is named as:
a) Finite population b) Infinite population
c) Real population d) None of these
9) Job tracker is created on
a )Namenode b) Datanode
c) Secondary datanode d) Secondary namenode
10) HDFS block size is larger as compared to the size of the disk blocks so that
a) Only HDFS files can be stored in the disk used.
b) The seek time is maximum
c) Transfer of a large files made of multiple disk blocks is not possible.
d) A single file larger than the disk size can be stored across many disks in the cluster
11) Which operating system is most Green?
a) Windows XP b) Windows Vista
c) Linux d) Apple OS Xs
12) Which of the following best describes predictive modeling?
a) A process marketers use to evaluate how factors influence future behavior
b) A process of building models that predict the future for businesses
c) A predictive analytics process that creates a statistical model of future behavior
d) None of these
13) chi-square is:
a) Statistical Modeling b) Probability Distribution
c) Set d) Distribution
14) State true or false : Velocity is an attribute of big data
a) True b) False
Q.2. A) Explain the following in short (Any Four) (8)
1) Reuse & Recycle
2) ITU-T
3) Types of Big Data
4) Business Intelligence
5) Training and Testing
6) Multitenancy
7)NIST
B) Write the answer of following (any Two) (6)
1) Explain Energy Saving Software Techniques.
2) Explain and compare traditional system vs Hadoop
3) Explain role of Data scientist
Q.3. A] Answer the following. (Any Two) (8)
1) Explain Population and Sampeling.
2) Define Hypervisor. Explain types and examples of Hypervisor.
3) What is Hadoop Architectue? Explain YARN component in details.
B] Answer the following. (Any One) (6)
1)Explain HDFS Architecture with suitable diagram
2) Explain types of virtualization.
Q.4. A] Answer the following. (Any Two) (10)
1) Explain Different frameworks available for IoTs
2)Explain characteristics of Machine Learning Tasks.
3) Explain Enterprise Green IT strategies.
B] Answer the following. (Any One) (04)
1) Define Hypervisor and its types.
2) Explain Statistical modeling and its types.
Q. 5) Write short note on (any Two) (14)
1) Explain IOT Architecture with diagram.
2)Explain Data Science Process.
3) Explain structure of Learning.
4) Explain V’s of Big Data.
SOLAPUR UNIVERSITY, SOLAPUR
BCA III (Semester V)(CBCS) Examination October 2019
Data Warehouse and Mining
Day & Date:
Time Duration: 2 Hr Total Marks: 70

Answer Key
Q. 1 A) MCQ
1 c 2 c
3 d 4 c
5 d 6 c
7 a 8 d
9 b 10 d
11 c 12 c
13 a 14 a

1)When you add a software stack, such as an applications to the service, the model shifts to
_____ model.
A. SaaS B. PaaS C. IaaS D. All of the mentioned
2) Which of the following is related to service provided by Cloud ?
A. Sourcing B. Ownership C. Reliability D. AaaS
3) Which of the following cloud concept is related to pooling and sharing of resources ?
A. Polymorphism B. Abstraction C. Virtualization D. None of the mentioned
4) Services to support DMBS in any cloud based system should be part of
A. SaaS B. PaaS C. Iaas D. None
5) How does Hadoop process large volumes of data?
A - Hadoop uses a lot of machines in parallel. This optimizes data processing.
B - Hadoop was specifically designed to process large amount of data by taking
advantage of MPP hardware.
C - Hadoop ships the code to the data instead of sending the data to the code.
D - Hadoop uses sophisticated caching techniques on name node to speed processing of data.
6) What is AVRO?
A. Avro is a data serialization library
B. Avro is a DATA compression library
C. Avro is a data library that create splittable files
D. None of these answers are correct
7) Which demon is responsible for replication of data in Hadoop?
A) HDFS B) Task Tracker C) Name Node D) Data Node
8) What are the five V’s of Big Data?
A. Volume B. Velocity C. Variety D. All the above
9) What is private cloud?A.
A standard cloud service offered via the !nternet
B. A cloud architecture maintained within an enterprise data center
C. A cloud service inaccessible to anyone but the cultural elite".
D. A cloud service rendered only for private organizations not public
10) KVM is type II______________________

Sr. Question and Options Correct


No. Option
1. What does “Velocity” in Big Data mean?
a) Speed of input data generation
b) Speed of individual machine processors ( D)
c) Speed of ONLY storing data
d) Speed of storing and processing data
2. What is Hive used as?
a) Hadoop query engine
b) MapReduce wrapper (D )
c) Hadoop SQL interface
d) All of the above
3. Which of the following are NOT true for Hadoop? (D)
a) It’s a tool for Big Data analysis
b) It supports structured and unstructured data
analysis
c) It aims for vertical scaling out/in scenarios
d) Both (a) and (c)
4. Hadoop is open source.
a) ALWAYS True
b) True only for Apache Hadoop ( B)
c) True only for Apache and Cloudera Hadoop
d) ALWAYS False
5. Which of the following are the core components of
Hadoop?
a) HDFS
( D)
b) Map Reduce
c) HBase
d) Both (a) and (b)
6. Hive can be used for real time queries.
a) TRUE
(B)
b) FALSE
c) True if data set is small
d) True for some distributions
7. What is the default HDFS block size? **
a) 32 MB
b) 64 KB (D)
c) 128 KB
d) 64 MB
8. What is the default HDFS replication factor? ( C)
a) 4
b) 1
c) 3
d) 2
9. Which of the following is NOT a type of metadata in
NameNode? **
a) List of files
( C)
b) Block locations of files
c) No. of file records
d) File access control information
10. NameNode tries to keep the first copy of data nearest
to the client machine.
a) ALWAYS true
b) ALWAYS False ( C)
c) True if the client machine is the part of the cluster
d) True if the client machine is not the part of the
cluster
11. NameNodes are usually high storage machines in the
clusters.
a) True
( B)
b) False
c) Depends on cluster size
d) True if co-located with Job tracker
12. Which of the following is the correct sequence of (C)
MapReduce flow?
a) Combine-> Reduce -> Map
b) Map->Reduce -> Combine
c) Reduce->Combine -> Map
d) Map-> Combine -> Reduce
13. Zookeeper ensures that
A - All the namenodes are actively serving the client
requests
B - Only one namenode is actively serving the client
requests
(B)
C - A failover is triggered when any of the datanode
fails.
D - A failover can not be started by hadoop
administrator.

14. The client reading the data from HDFS filesystem in


Hadoop
A - gets the data from the namenode
B - gets the block location from the datanode (c)
C - gets only the block locations form the namenode
D - gets both the data and block location from the
namenode
15. The hadoop frame work is written in
A - C++
B - Python (c)
C - Java
D - GO
16. Which of the following are among the duties of the (D)
Data Nodes in HDFS?
A - Maintain the file system tree and metadata for all
files and directories.
B - None of the options is correct.
C - Control the execution of an individual map task
or a reduce task.
D - Store and retrieve blocks when told to by clients
or the NameNode.
17. Hadoop is a framework that works with a variety of
related tools. Common cohorts include
____________
a)MapReduce, Hive and HBase (a)
b)MapReduce, MySQL and Google Apps
c)MapReduce, Hummer and Iguana
d) MapReduce, Heron and Trumpet
18. What was Hadoop named after?
a) Creator Doug Cutting’s favorite circus act
b) Cutting’s high school rock band
(c)
c) The toy elephant of Cutting’s son
d) A sound Cutting’s laptop made during Hadoop
development
19. All of the following accurately describe Hadoop, (b)
EXCEPT ____________
a) Open-source
b) Real-time
c) Java-based
d) Distributed computing approach
20. __________ has the world’s largest Hadoop cluster.
a) Apple
b) Datamatics (c)
c) Facebook
d) None of the mentioned
21. Which of the following is the daemon of Hadoop?
a) NameNode
b)Node manager (d)
c) DataNode
d) All of the above
22. Hadoop works in
a) master-worker fashion
b)master – slave fashion (b)
c)worker/slave fashion
d)All of the mentioned
23. As compared to RDBMS, Apache Hadoop
a) Has higher data Integrity
b) Does ACID transactions
(d)
c) Is suitable for read and write many times
d) Works better on unstructured and semi-structured
data.
24. Which of the following is the true about metadata? (c)
A. Metadata shows the structure of HDFS
directories/files
B. FsImage & EditLogs are metadata files
C. Metadata contain information like number of
blocks, their location, replicaswrong
D. All of the above
25. Which one of the following stores data?
 A. Name node
 B. Data nodecorrect
(B)
 C. Master node
 D. None of these

26. Can multiple clients write into an HDFS file


concurrently?
a) Yes (b)
b) No

27. Is it possible to provide multiple inputs to Hadoop?


a) Yes
(a)
b) No

28. Pig is a: (b)


a) Programming Language
b) Data Flow Language
c) Query Language
d) Database
29. Which one of the following options define the role of
job tracker in big data?
a) Resource management
b) Clustering (a)
c) Task execution
d) None of the above

30. Which one of the following option is not part of the


five Vs of big data?
a) Volume
b) Variety (c)
c) Vitality
d) Veracity

31. Sources of Big data


A. Web server logs
B. Social madia Activity reports (D)
C. Information captured by sensors
D. All the three
32. What is the name of the programming framework
originally developed by Google that supports the
development of applications for processing large data
sets in a distributed computing environment?
a) MapReduce
b) Hive
c) ZooKeeper

You might also like