INSTALLATION DOCUMENTS BY RAVI

Friday, April 27, 2018

obiee 10.1.3.4 installation on windows

Right click on the setup file and select run as administrator






































Click Next

















Provide installation location, data location and installation type

Click Next

















Choose the setup type as complete and click next

















Provide java home and administrator password and click next

















Provide local system account details and click next

















Select the language and click next

















Review the summary and click next

















Wait until installation process completes

































Click Next

















Review the installation summary and click next

















Click Next

We must restart the the system to complete the installation successfully.

After reboot start obiee as below


























Attunity Replicate server installation and configuration


Automating and accelerating data pipelines for data lakes and the cloud
Attunity is changing data integration with the latest release of the Attunity platform to deliver data, ready for analytics, to diverse platforms on-premises and in the Cloud. 
Unlike the traditional batch-oriented and inflexible ETL approaches of the last decade, Attunity provides the modern, real-time architecture you need to harness the agility and efficiencies of new data lakes and cloud offerings.

Streaming:
Universal Stream Generation
Databases can now publish events to all major streaming services, including Kafka, Confluent, Amazon Kinesis, Microsoft Azure Event Hub, MapR Streams.

Optimized Data Streaming
Flexible message formats including JSON and AVRO along with the separation of data and metadata into separate topics allows for smaller data messages and easier integration of metadata into various schema registries.

Cloud:
AWS S3 and Kinesis
Data can now land in S3 either in bulk load, change data capture, or published to Amazon Kinesis

Snowflake Data Loading
The entire catalog of Attunity supported data sources can now feed a Snowflake data warehouse in bulk load or via change data capture.

Preferred data movement solution for Amazon Web Services and Microsoft Azure
Deep partnerships and broad product integration with industry leaders

Data Lakes:
Automate the Creation of Analytics-Ready Data Lakes
Data lands in optimized time based partitions and Attunity automatically creates the schema and structures in the Hive Catalog for Operational Data Stores (ODS) and Historical Data Stores (HDS) – with no manual coding. Supported distributions include Hortonworks, AWS Elastic MapReduce, and Cloudera.

Ensure Data Consistency
Attunity automatically reconciles continuous data inserts, updates and deletions, while providing ACID compliance, without manual coding or disruption. Attunity also recognizes and responds to source data structure changes (DDL) and automatically applies changes to your data lake.

Enterprise Management & Control:
Enterprise-wide Control and Management
Scalability to thousands of tasks; resiliency and recovery to maintain data integration processes across multiple data centers and hybrid cloud environments

Performance and Data Flow Analytics
Comprehensive historical and real-time reporting for improved capacity planning and performance monitoring of all data flows

Operational Metadata Creation and Discovery
Central repository shared across the Attunity platform and with third party tools for enterprise-wide reporting

Micro services API
New REST and .NET APIs designed for invoking and managing
Attunity services using a standard web-based UI

Downloading the software:
https://4web.s3.amazonaws.com/Files/Replicate/AttunityReplicate_Express_Linux_X64.rpm

Installing Attunity

Prerequisite
1. Windows or Linux 64-bit Server (dedicated to running 'Attunity Replicate' software)
  • Located in same local network as Source Database(s)
  • Will host 'Attunity Replicate' software (download available via Welcome Email sent automatically when one activates the AMI)
    • Minimum Hardware Requirements:
    • Windows Server 2008 R2, 2012 or 2012 R2
    • Red Hat Linux Enterprise Linux 6.2 and above
    • SUSE Linux 11 and above
    • Quad-Core processor, 8 GB RAM, and 320 GB of disk space.

2. Attunity Replicate Console is web-based and requires one of the following browsers:
  • Microsoft Internet Explorer Version 9 or higher
  • Mozilla Firefox Version 38 and above
  • Google Chrome


Installation:

Run the downloaded rpm "AttunityReplicate_Express_Linux_X64.rpm" as below








Verifying that the Attunity Replicate Server is Running:








Accessing the Attunity Replicate Express Console:
Attunity Replicate Server on Windows:
https://<computer name>/AttunityReplicate
Attunity Replicate Server on Linux:
https://<computer name>:<port>/AttunityReplicate
Where <computer name> is the name or IP address of the computer where the Attunity Replicate Server is installed and <port> is the C UI Server port (3552 by default).
Open browser and type below url
https://localhost:3552/AttunityReplicate

Setting console password:






Set the password as below and restart the server









Open the console now























Configuring Source:

Open the 'Attunity Replicate Console' and select 'Manage Endpoint Connections'
























Click ‘New Endpoint Connection'























Configure with 'Add Database' with these values:
























Test the connection and click on save



Configuring Target:

Open the 'Attunity Replicate Console' and select 'Manage Endpoint Connections'
























Click ‘New Endpoint Connection'






















Configure with 'File' with these values:
























Test the connection and click on save
























Creating, Running, and Monitoring Replicate Tasks

1: Open Attunity Replicate and select 'New Task'. In the 'New Task' window, enter a unique name, then select 'OK'.














































2: Drag your recently configured source database to the 'Drop source database here' area on the right.
3: Drag your recently configured target database to the 'Drop target database here' area on the right




























4: Select 'Table Selection' on the right. The 'Select Tables' window opens.
5: From the 'Schema' drop-down list, select a schema.
6: Click 'Search' to find the tables in the Schema. You may then 'add' / 'remove' / 'add all' / 'remove all'. Then click 'OK'.













































































Note: As this is trail version all the features are not available. Attunity Replicate is not an open source software.


Unauthorized connection for super-user: root from IP 127.0.0.1

When trying to upload files to hdfs through ambari console getting the below error

Unauthorized connection for super-user: root from IP 127.0.0.1
























Looks like we have not added the mentioned IP Address properly inside the property: hadoop.proxyuser.root.hosts

If ambari-server daemon is running as root, we need to set up a proxy user for root in core-site by adding and changing properties in HDFS > Configs > Custom core-site:
  1. hadoop.proxyuser.root.groups=*
  2. hadoop.proxyuser.root.hosts=*
Open ambari console


Go to HDFS -> Configs























Click on Custom core-site























Modify the below parameters























Save the configuration











Click ok


Restart the affected components























After successful restart try to upload files now























How to get the windows operating system version from command prompt


Open command prompt







































Navigate to C:\Windows\System32















Run winver.exe to get the version of operating system




































Thursday, April 26, 2018

ERROR: Attempting to operate on hdfs datanode as root

Getting error when we tried to start hadoop as a root user












To resolve this navigate to /usr/hdf/3.1.1.0-35/hadoop/etc/hadoop
open hadoop-env.sh file
Add the below lines
export HDFS_NAMENODE_USER="root"
export HDFS_DATANODE_USER="root"
export HDFS_SECONDARYNAMENODE_USER="root"
export YARN_RESOURCEMANAGER_USER="root"
export YARN_NODEMANAGER_USER="root"





















Start hadoop now




























Purge Oracle Analytics Cloud cache using curl command

Clearing cache for all databases: Syntax: curl -i -X DELETE -u joe@example.com:mypassword1 -H "X-ID-TENANT-NAME:myidentitydomain...