C2090-102 Exam - IBM Big Data Architect

certleader.com

Proper study guides for Up to the immediate present IBM IBM Big Data Architect certified begins with IBM C2090-102 preparation products which designed to deliver the Pinpoint C2090-102 questions by making you pass the C2090-102 test at your first time. Try the free C2090-102 demo right now.

Free demo questions for IBM C2090-102 Exam Dumps Below:

NEW QUESTION 1
You are going to develop Spark applications for your client. Which of the following cluster managers, that is also part of the Open Data Platform (ODP), would you use to focus on building the business driven application?

  • A. Amazon EC2
  • B. Apache Mesos
  • C. Hadoop Yarn
  • D. Standalone deploy mode

Answer: C

NEW QUESTION 2
In designing a new Hadoop system for a customer, the option of using SAN versus DAS was brought up. Which of the following would justify choosing SAN storage?

  • A. SAN storage provides better performance than DAS
  • B. SAN storage reduces and removes a lot of the HDFS complexity and management issues
  • C. SAN storage removes the Single Point of Failure for the NameNode
  • D. SAN storage supports replication, reducing the need for 3-way replication

Answer: D

NEW QUESTION 3
Company A is searching for a browser-based visualization tool to perform analysis
on vast amounts of data in any structure. They want to execute operations such as pivot, slice and dice, among others. Which of the following would meet these requirements?

  • A. Streams
  • B. BigSheets
  • C. Aginity Workbench
  • D. Watson Explorer

Answer: B

Explanation:
References:
http://www.dotgroup.co.uk/wp-content/uploads/2014/11/Harness-the-Power-of-Big-Data- The-IBM-Big-DataPlatform.pdf Page: 132

NEW QUESTION 4
Data stored in Netezza can be used for model building, scoring, and model refresh. How is SPSS Modeler connectivity to Netezza enabled?

  • A. Open Database Connectivity (ODBC) driver provided by Netezza
  • B. DataDirect Connect for ODBC drivers
  • C. OLE DB drivers provided by Netezza
  • D. Java Database Connectivity (JDBC) driver provided by Netezza

Answer: A

Explanation:
Reference: http://www.ibm.com/developerworks/data/library/cognos/infrastructure/cognos_sp ecific/pag e569.html

NEW QUESTION 5
As data architect, understanding where patient data resides, how is it related, and how you define security and privacy policies requires knowledge of each of the following?

  • A. PII Data
  • B. HIPAA Requirements
  • C. Privacy by design
  • D. Data storage in HDFS

Answer: B

NEW QUESTION 6
You are building a pricing application combining the output from your internal customer database along with an online feed that contains up to the minute competitor pricing adjustments. Which of the following is the best choice for data level real time output?

  • A. Big R
  • B. Big SQL
  • C. Hadoop native hdfs map reduce based queries
  • D. Infosphere streams

Answer: D

Explanation:
References:
http://www-03.ibm.com/software/products/en/ibm-streams

NEW QUESTION 7
A reputable market research firm wants to explore more business opportunities. They have great in house skill in python and machine learning. Their business model is simple, they build the solutions for customers using python and machine learning algorithms and give these solutions to the customer’s engineering team for implementation. Given this scenario, which of the following would you recommend?

  • A. Netezza
  • B. Spark
  • C. Cloudant
  • D. Hadoop

Answer: D

NEW QUESTION 8
Which one of the following statements is TRUE?

  • A. Big SQL uses Hadoop MR framework to process query tasks in parallel
  • B. Big SQL executes queries locally on Big SQL server single node on a multi node cluster
  • C. Big SQL can process queries in parallel and executes queries locally
  • D. Big SQL only works with HDFS

Answer: A

Explanation:
Reference:
https://www- 01.ibm.com/support/knowledgecenter/SSPT3X_3.0.0/com.ibm.swg.im.infosphere.biginsight s.analyze.doc/doc/bigsql_performance.html

NEW QUESTION 9
Which of the following statements is TRUE regarding Cloud deployment models?

  • A. Performance and scalability requirements are a critical factor for deciding between Platform as a Serviceand Infrastructure as a Service deployment models
  • B. In a platform as a Service offering, the customer has root access to the servers
  • C. Applications with extremely high transactions volumes are good candidates for Platform as a Service
  • D. In an infrastructure as a service deployment, the cloud provider provides security patching, monitoring andfail over capabilities

Answer: A

NEW QUESTION 10
If the recovery point objective (RPO) is low, which of the following techniques would be the most appropriate?

  • A. Clustering
  • B. Database Shadowing
  • C. Remote Journaling
  • D. Tape Backup

Answer: A

Explanation:
References:
http://whatis.techtarget.com/definition/recovery-point-objective-RPO

NEW QUESTION 11
A component of IBM Industry Model forms the basis of the “Logical Data Warehouse Model” that spans across the traditional RDBMS and Hadoop technology. It defines all of the data structures that would be expected to be defined in the Detailed System of Record. What is the name of this component?

  • A. Business data model
  • B. Atomic warehouse model
  • C. Dimensional warehouse models
  • D. Metadata management

Answer: A

Explanation:
C2090-102 dumps exhibitReferences:
http://www.ibm.com/support/knowledgecenter/SS9NBR_9.1.0/com.ibm.ima.using/comp/bdm/intro.dita

NEW QUESTION 12
A large application vendor wants to port their existing distributed applications to run on Hadoop. In order to be competitive they need to provide monitoring and keep the size of the monitored applications consistent with the configuration. This implies the ability to deploy a replacement, for example, for any failed components. Which of the following would be a workable solution?

  • A. Nagios with YARN
  • B. Slider with YARN
  • C. Oozie with Lucene
  • D. OPTIM Performance Manager

Answer: B

NEW QUESTION 13
Which of the following statements is TRUE regarding cloud applications?

  • A. Migrating a legacy application to the cloud is a simple solution to drive down cost
  • B. Architecting and deploying a scalable cloud application requires a private cloud implementation
  • C. To be truly elastic, a cloud application must be tightly coupled to the physical cloud environment
  • D. Leveraging a private v
  • E. public cloud may result in sacrificing some of the core advantages of cloud computing

Answer: B

NEW QUESTION 14
The AQL query language is the easiest and most flexible tool to pull structured output from which of the following?

  • A. Hive data structures
  • B. Unstructured text
  • C. Hbase schemas
  • D. JDBC connected relational data marts

Answer: A

Explanation:
Reference:
http://www.ibm.com/developerworks/library/bd-sqltohadoop2/

NEW QUESTION 15
What are the two levels documented in the Operational Model? (Choose two.)

  • A. Logical
  • B. Rational
  • C. Theoretical
  • D. Physical
  • E. Middleware

Answer: AC

NEW QUESTION 16
You need to provision a Hadoop cluster to perform data analysis on customer sales data to predict which products are more popular. Which of the following solutions will let you set up your cluster with the most stability in the platform?

  • A. Purchase specific products from multiple Independent Software Vendors (ISV) for your requirements inorder to take advantage of vendor-specific features
  • B. Develop your own platform of software components to allow for maximum customization
  • C. Use a Hybrid of ISV applications to build your customizations on top of that
  • D. Leverage the Open Data Platform (ODP) core to provide a stable base against which Big Data solutionsproviders can qualify solutions

Answer: D

Explanation:
References:
http://hortonworks.com/innovation/open-data-platform/

NEW QUESTION 17
Which of the following can NOT be performed by the Open Data Platform (ODP)?

  • A. Provide development and governance of upstream projects
  • B. Contribute to the Apache Software Foundation projects
  • C. Produce a set of tools and methods that enable members to create and test differentiated offerings based on the ODP core
  • D. Accelerate the delivery of Big Data solutions by providing a well-defined core consisting of all the components of the Apache Software Foundation projects

Answer: C

NEW QUESTION 18
What is the most important aspect of data center disaster recovery?

  • A. A complete damage assessment
  • B. Control of critical assets
  • C. Restoration of business functions
  • D. Protection of individual life

Answer: D

NEW QUESTION 19
The analysis layer reads the data digested by the layer massaging and store layer. In some cases, the analysis layer accesses the data directly from the data source. Designing the analysis layer requires careful forethought and planning. Decisions must be made with regard to how to manage the tasks to do which of the following?

  • A. Reduce the volume of data
  • B. Locate the data sources that can provide data for the required entities
  • C. Identify multi-dimension data sources
  • D. Identify data import and export tools to adopt and deliver an ETL strategy

Answer: B

Explanation:
References:
http://www.ibm.com/developerworks/library/bd-archpatterns3/

NEW QUESTION 20
In a typical Hadoop HA cluster, two separate machines are configured as which of the following?

  • A. Data Nodes
  • B. Edge Nodes
  • C. Name Nodes
  • D. None of the Above

Answer: A

Explanation:
Reference:
http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html

NEW QUESTION 21
For company B, 85% of their analytics queries only involve about 25% of their data; another 10% of the queries will touch 35% of the rest of the data, and only 5% of the queries will touch the remaining 40% of the data. The estimated volume is 50TB growing at 1 TB per year. Which of the following would provide the best value (business benefit) and lowest TCO?

  • A. Place the entire set of data in a data warehouse with proper partitioning and indexing
  • B. Place the entire set of data in a hadoop environment – using commodity HW
  • C. Place the top 25% of data (used by 85% of the query) in a hadoop environment, and the rest in a data warehouse
  • D. Place the top 25% of data (used by 85% of the query) in a data warehouse, and the rest in a hadoop environment

Answer: C

NEW QUESTION 22
......

100% Valid and Newest Version C2090-102 Questions & Answers shared by Dumps-files.com, Get Full Dumps HERE: https://www.dumps-files.com/files/C2090-102/ (New 110 Q&As)