AWS BDS-C00 Certified Big Data Speciality Practice Test Set 1

You need to perform ad-hoc analysis on log data, including searching quickly for specific error codes and reference numbers. Which should you evaluate first?


Options are :

  • AWS Elastic search Service (Correct)
  • AWS EMR
  • AWS Red shift
  • AWS Dynamo DB

Answer : AWS Elastic search Service

Which of the following node types are recommended when launching Red shift cluster ?


Options are :

  • General Purpose
  • Memory Optimized
  • Dense Storage (Correct)
  • Accelerated Computing

Answer : Dense Storage

Your red shift cluster has 2 dsl .XL nodes. How many files would you normally split your data into for loading Into the Red shift cluster Please select:


Options are :

  • 4 (Correct)
  • 5
  • 10
  • 2

Answer : 4

You need to filter and transform incoming messages coming from a smart sensor you have connected with AWS. Once messages are received, you need to store them as time series data in Dynamo DB. Which AWS service can you use?


Options are :

  • loT Rules Engine (Correct)
  • loT Device Shadow
  • loT Message broker
  • loT Device Shadow service

Answer : loT Rules Engine

You need to store a large volume of data. The data needs to be readily accessible for a short period, but then needs to be archived indefinitely after that. What is a cost-effective solution that can help fulfill this requirement?


Options are :

  • Store your data in Amazon 53. and use lifecycle policies to archive to Amazon Glacier (Correct)
  • Keep all your data in 53 since this is a durable storage
  • Store your data in an EBS volume and use lifecycle policies to archive to Amazon G?acier
  • Store your data in Amazon 53, and use lifecycle policies to archive to S3-lnfrequently Access

Answer : Store your data in Amazon 53. and use lifecycle policies to archive to Amazon Glacier

Which of the below components of a Red shift cluster, if down, can render the Red shift cluster as unavailable? Please select:


Options are :

  • Leader Node (Correct)
  • Master Node
  • Core Node
  • Compute Node

Answer : Leader Node

What is the term given to a group of data records in a stream in AWS Kinesis?


Options are :

  • Group Stream
  • Consumer
  • Batch
  • Shard (Correct)

Answer : Shard

You need to use AWS services which would meet the following requirements a. Ability to load streaming data b. Ability to have scalable private connections to on-premise data centers Which of the below mentioned services would meet this requirement. Choose 2 answers from the options given below Please select:


Options are :

  • Direct Connect (Correct)
  • Data Migration Services
  • Snowball
  • Kinesis Fire hose (Correct)

Answer : Direct Connect Kinesis Fire hose

Which of the following can be used to monitor EMR Clusters and give reports of the performance of the cluster as a whole?


Options are :

  • Ganglia
  • Cloudwatch logs (Correct)
  • Cloud trail
  • AWS Config

Answer : Cloudwatch logs

There is a requirement to convert and migrate an on-premise Oracle database to AWS Aurora. Which of the following make up the steps that are involved in the process? Choose 3 answers from the options given below


Options are :

  • Convert database schema and code using AWS Schema Conversion Tool (Correct)
  • Use AWS Data Pipeline to transfer the data from Oracle to AWS Aurora
  • Migrate data from the source database to the target database using AWS Database Migration Service (Correct)
  • Perform post-migration activities (Correct)

Answer : Convert database schema and code using AWS Schema Conversion Tool Migrate data from the source database to the target database using AWS Database Migration Service Perform post-migration activities

You need real-time reporting on terabyte-scale of log data that are being generated from your application. In addition, you need anomaly detection. The processing latency needs to as less as possible. Out of the below options, choose 2 answers that would be able to fulfill this requirement?


Options are :

  • AWS Fire hose
  • Apocne Sp3r
  • Spark MLlio (Correct)
  • Amazon ML

Answer : Spark MLlio

Which of the options is best for an interactive and collaborative notebook for data exploration?


Options are :

  • D3
  • Zeppelin (Correct)
  • Kinesis Analytics
  • Hive

Answer : Zeppelin

You are building a game high score table in Dynamo DB. You will store each user?s highest score for each game. with many games, all of which have relatively similar usage levels and numbers of players. You need to be able to look up the highest score for any game. What?s the best Dynamo DB key structure?


Options are :

  • Game ID as the hash key. Highest Score as the range key. (Correct)
  • Highest Score as the hash I only key.
  • Game ID as the hash / only key.
  • Game ID as the range I only key

Answer : Game ID as the hash key. Highest Score as the range key.

Which of the following storage types can be used with Amazon EMR. Choose 3 answers from the options given below?


Options are :

  • Local file system (Correct)
  • HDFS (Correct)
  • EMRFS (Correct)
  • Amazon Glacier

Answer : Local file system HDFS EMRFS

You to create an Amazon Machine Learning model to predict how many inches of snow will fall in an area based on the historical snowfall data. What type of modeling will you use?


Options are :

  • Regression (Correct)
  • Categorical
  • Simple
  • Binary

Answer : Regression

There is a requirement to analyze a large set of data updates from Kinesis and Dynamo DB. Which of the following tools can be used for this purpose.?


Options are :

  • Elastic search (Correct)
  • EMR
  • Quick sight
  • Reds shift

Answer : Elastic search

You need services can be used to quickly and cost-effectively perform Extract Transform Load (ETL) on large datasets. The data source Is from AWS 53 and the processed results are also put into 53. Which of the below services can be used?


Options are :

  • AWS (Correct)
  • AWS SOS A
  • AWS LOT
  • AWS Kinesis

Answer : AWS

You work for a well establish web-based advertising company which has set of predefined ads rotating displayed routinely. Your website is getting more and more popular with high diverse visitors. You implemented a own solution to place dynamic ads based on the user click behavior or pattern analysis. Later you discovered the process time is not keeping up to display the new ads since most users just stay at your site for seconds not minutes and could navigate out at any time. You start evaluating AWS services and features for possible solution. Which of the below is your recommendation?


Options are :

  • Push web clicks to Amazon Kinesis Stream and analyze with kinesis analytics or Kinesis Client Library (KCL) (Correct)
  • You decide to move your website to host in AWS to use AWS Kinesis to dynamically process user click stream in real-time.
  • Push the click stream data to Amazon SQS queue which is subscribed by your application and send to Amazon RDS and analyze with SQL
  • Push web clicks to Amazon Kinesis Fire hose and analyze with kinesis analytics or Kinesis Client Library You have just created a Red shift cluster in AWS.

Answer : Push web clicks to Amazon Kinesis Stream and analyze with kinesis analytics or Kinesis Client Library (KCL)

There is a requirement to perform SQL querying along with complex queries on different backend data that include Red shift, My SQL Hive on EMR. H3, and PostgreSQL Which of the below tool can run queries on all the different platforms for your daily adhoc analysis? Please select:


Options are :

  • Presto S (Correct)
  • Quick Sight
  • EMR
  • Athena

Answer : Presto S

Which of the following of Dynamo DB is an index with a partition key and a sort key that can be different from those on the base table Please select:


Options are :

  • Global Primary Index
  • Local Secondary Index
  • Local Primary Index (Correct)
  • Global Secondary Index

Answer : Local Primary Index

Which of the following in a Dynamo DB-compatible caching service?


Options are :

  • Redis
  • Mem cache
  • DAX (Correct)
  • Elastic cache

Answer : DAX

You are in the process of designing a system which is going to make use of the EMR service. The EMR system will be used to process highly confidential data. Which of the following options can be used to encrypt data at rest. Choose 2 answers from the options given below.


Options are :

  • TLS
  • 553-KMS (Correct)
  • LUKS
  • SSL

Answer : 553-KMS

You run a data processing EMR Hadoop job once for several hours a day then terminate it after job done. \The EMR cluster is running in AWS that uses EMRFS on S3 for the data. The data in S3 is constantly changing and you need to ensure the updated data Is reflected In EMRFS? You also want to the keep the data after the EMR cluster terminated. How can this be achieved?


Options are :

  • Enable versioning on the S3 bucket.
  • Consider using EBS Volumes to store the data instead.
  • Enable recycling on the EMR cluster
  • Consider using Instance storage with cluster termination protection.
  • Enable Consistent Views (Correct)

Answer : Enable Consistent Views

Which of the following commands can be used to transfer data from Dynamo DB to Red shift?


Options are :

  • UNLOAD
  • Dist Cp
  • EXPORT (Correct)
  • COPY

Answer : EXPORT

You need to load a lot of data once a week from your on-premise datacenter on to AWS Redshift. Which of the below AWS-Managed Cloud Data Migration Tools can be used for this data transfer in simple, fast, and secure way. Choose 2 answers from the options given below. Please select:


Options are :

  • Data Pipeline
  • Import/Export to AWS
  • Snowball
  • Direct Connect (Correct)

Answer : Direct Connect

You need a cost-effective solution to store a large collection of video files and have fully managed data warehouse service that can keep track of and analyze all your data efficiently using your existing business intelligence tools. Which of the following would form the solution required to full fill the requirements?


Options are :

  • Store the data in Amazon S3 and reference its location in Amazon EMR. Amazon EMR will keep track of metadata about your video files, but the video files themselves would be stored in Amazon 53.
  • Store the data in Amazon $3 and reference its location in Amazon Dynamo DB. Amazon Dynamo DB will keep track of metadata about your binary objects, but the large objects themselves would be stored in Amazon S3.
  • Store the data in Amazon Dynamo DB and reference its location in Amazon Red shift. Amazon Red shift wil keep track of metadata about your video files, but the video files themselves would be stored in Amazon S3.
  • Store the data in Amazon S3 and reference its location in Amazon Red shift. Amazon Red shift will keep track of metadata about your binary objects. but the large objects themselves would be stored in Amazon 53. (Correct)

Answer : Store the data in Amazon S3 and reference its location in Amazon Red shift. Amazon Red shift will keep track of metadata about your binary objects. but the large objects themselves would be stored in Amazon 53.

You have an ecommerce application hosted in AWS. This application sells items online to customers. You want to use the Amazon Machine Learning Model to see the probability of a person buying a particular product. Which of the following ML model can be used to predict, whether a person will buy a specific product or will not buy it?


Options are :

  • Binary (Correct)
  • Simple
  • Regression
  • Categorical

Answer : Binary

You currently work for a company that looks at baggage handling. There are GPS devices located on the baggage delivery units to deliver the coordinates of the unit every 10 seconds. You need to process these coordinates in real-time from multiple sources. Which tool should you use to digest the data


Options are :

  • AWS Data Pipeline
  • Amazon SQS
  • Amazon EMR
  • Amazon Kinesis (Correct)

Answer : Amazon Kinesis

Which of the below mechanisms can be for authentication of loT devices for AWS loT? Select 3 correct options ?


Options are :

  • Amazon Cognito identities (Correct)
  • KMS
  • lAM users. groups. and roles (Correct)
  • X.509 certificates

Answer : Amazon Cognito identities lAM users. groups. and roles

Your application generates a 1 KBJSON payload that needs to be queued and delivered EC2 instances for applications. At the end of the day, the application needs to replay the data for the past 24 hours, Which of the following services would you use for this requirement?


Options are :

  • SNS
  • SQS
  • Kinesis (Correct)
  • Kinesis Fire hose

Answer : Kinesis

Comment / Suggestion Section
Point our Mistakes and Post Your Suggestions