tibco activematrix businessworks and big data
TRANSCRIPT
![Page 1: TIBCO ActiveMatrix BusinessWorks and Big Data](https://reader036.vdocuments.mx/reader036/viewer/2022083018/577c7ec11a28abe054a250d6/html5/thumbnails/1.jpg)
TIBCO ActiveMatrix BusinessWorks and Big Data.
This article gives an overview of big data and TIBCO ActiveMatrix BusinessWorks Plug-in for Big Data.
![Page 2: TIBCO ActiveMatrix BusinessWorks and Big Data](https://reader036.vdocuments.mx/reader036/viewer/2022083018/577c7ec11a28abe054a250d6/html5/thumbnails/2.jpg)
TIBCO ActiveMatrix BusinessWorks Plug-in
• TIBCO BusinessWorks is an easy-to-use integration software that provides
• world-class integration technology in a rapidly deployable solution that manages
• the entire life cycle of integration projects.• TIBCO ActiveMatrix BusinessWorks Plug-in for Big Data plugged
into TIBCO• BusinessWorks, allows you to use the established non-code
approach to integrate• with Hadoop family projects, such as Hadoop Distributed File
System (HDFS).
• **Before running operations by using this plug-in, ensure that the user hasappropriate permissions on HDFS server.
![Page 3: TIBCO ActiveMatrix BusinessWorks and Big Data](https://reader036.vdocuments.mx/reader036/viewer/2022083018/577c7ec11a28abe054a250d6/html5/thumbnails/3.jpg)
HDFS Palette and Hadoop Palette
• The plug-in contains the following two palettes:• • The HDFS palette allows you to do some operations on the files in HDFS.• — The HDFSConnection shared resource contains all necessary parameters to• connect to HDFS.• — The HDFS Operation activity is used to do basic operations on files in• HDFS.• — The ListFileStatus activity is used to list the status of the specified file or• directory.• — The Read activity is used to read data from a file in HDFS.• — The Write activity is used to write data to a file in HDFS.• • The Hadoop palette allows you to utilize the benefits of Hive, Mapreduce, and• Pig based on Hadoop.• — The HCatalogConnection shared resource contains all the necessary• parameters to connect to HCatalog.• — The Database shared resource is an administrative container for a set of• tables.• — The Table shared resource provides shared virtual storage for data.• — The Hive activity is used to facilitate querying and managing large datasets• residing in distributed storage.• — The Mapreduce activity is used to create and queue a standard Mapreduce• job or a streaming Mapreduce job.• — The Pig activity is used to create and queue a Pig job.• — The WaitForJobCompletion activity is used to wait for the specified job to• complete until it reaches the specified value of timeout.
![Page 4: TIBCO ActiveMatrix BusinessWorks and Big Data](https://reader036.vdocuments.mx/reader036/viewer/2022083018/577c7ec11a28abe054a250d6/html5/thumbnails/4.jpg)
• The HDFS palette includes one shared resource and four activities, as shown• in Figure 1.
• Shared resource:• — HDFSConnection• Activities:• — HDFS Operation• — ListFileStatus• — Read• — Write• • The Hadoop palette includes three shared resources and four activities, as• shown in Figure 2.• Figure 2 HCatalog Palette• TIBCO ActiveMatrix BusinessWorks Plug-in for Big Data User’s Guide• Palette Overview | 5• Shared resources:• — HCatalogConnection• — Database• — Table• Activities:• — Hive• — Mapreduce• — Pig• — WaitForJobCompletion