Questions tagged [cloudera-cdp]

Following are the CDP's key feature's:-

  • On-premises and public cloud
  • Multi-cloud and multi-function
  • Simple to use and secure by design
  • Manual and automated
  • Open and extensible
  • For data engineers and data scientists

Information on CDP Private Cloud is here Information on CDP Public Cloud is here

For free training on CDP register here

CDP Private Cloud Overview

CDP Public Cloud Overview

Related tags:-

28 questions
2
votes
0 answers

How do I Create Hive External table on top of ECS S3 object storage using "S3a//" protocol

I am trying to create Hive external table using Beeline on top of S3 object storage using "S3a//" scheme.I have followed the official cloudera documentation and configured the below properties. fs.s3a.access.key fs.s3a.secret.key fs.s3a.endpoint I…
1
vote
0 answers

In CDP how to update OneViewofProfile Id through (VisitorId, browserID)?

anyone from CDP certified guide me on this. what is the better approach to update CDP profile id, and how these below use cases are same. When 2 user are using same device, do they got the same visitor Id ? One CDP platform I visit there i…
1
vote
1 answer

Hive managed table issue to create a hive table from a hdfs location in CDP

I have a CDP 7.3.1 where using sqoop , I have loaded data from Postgres database table into HDFS location /ts/gp/node. Now I am trying to create a hive table on this. I get the below error. Please help CREATE TABLE dsk.node ( district string, …
stacktesting
  • 103
  • 8
1
vote
1 answer

Scala - How to read MQ message which exceed 4096 characters

Application Information: IBM MQ 9.2, Cloudera CDP 7.1.6, Spark 2.4.5 I am upgrading the spark code from Spark 1.6 to Spark 2.4.5. I have a json content (complex schema) push to the MQ Queue which the message length exceed 4096. I able to read the…
Chia
  • 11
  • 2
1
vote
1 answer

Unable to create Managed Hive Table after Hortonworks (HDP) to Cloudera (CDP) migration

We are testing our Hadoop applications as part of migrating from Hortonworks Data Platform (HDP v3.x) to Cloudera Data Platform (CDP) version 7.1. While testing, we found below issue while trying to create Managed Hive Table. Please advise on…
Vasanth Subramanian
  • 1,040
  • 1
  • 13
  • 32
1
vote
1 answer

How to migrate roles from one apache ranger instance to other instance?

We are planning to make a replica cluster of existing CDP cluster. I can import/export policies but can not import/export roles. We have around 2k+ roles, using following api I can create role but wanted to know if this can be…
potatoaim
  • 75
  • 1
  • 1
  • 5
1
vote
1 answer

Read/Write with Nifi to Kafka in Cloudera Data Platform CDP public cloud

Nifi and Kafka are now both available in Cloudera Data Platform, CDP public cloud. Nifi is great at talking to everything and Kafka is a mainstream message bus, I just wondered: What are the minimal steps needed to Produce/Consume data to Kafka from…
Dennis Jaheruddin
  • 21,208
  • 8
  • 66
  • 122
0
votes
1 answer

How to find time difference between two timestamps in seconds and milliseconds in hive and impala

Need a help in finding time difference between two timestamps in seconds and milliseconds in hive and impala. We are using CDP cluster. Two columns are in string datatype with value in the format…
0
votes
0 answers

Hue Pyspark connector using Livy - Increate spark driver memory for interactive sessions

We are using CDP private cloud 7.1.7 and have configured Hue connector for pyspark using livy. By default I can see the driver launches with 1GB memory and I need to increase this as some of the code we are running is failing due to OOM errors for…
0
votes
0 answers

Sqoop export geometry column : ERROR orm.ClassWriter: No Java type for SQL type 1111 (postgresql geometry point 3945)

Using Sqoop export on CDP cluster (hdfs), I'm trying to export data geometry to Postgresql table but I'm facing a Sqoop ERROR orm.ClassWriter: No Java type for SQL type 1111 for column 'geom'. The 'geom' column definition on PostGis is : geometry…
Darko37
  • 25
  • 5
0
votes
0 answers

Unable to Connect to Hive using jdbc with CDP version

Driver: com.cloudera.hive.jdbc41.HS2Driver jdbc url:  jdbcUrl=jdbc:hive2://EXAMPLE.NET:10010/;AuthMech=1;KrbHostFQDN=EXAMPLE.NET/;KrbServiceName=hive;krbRealm=REALM;SSL=1;SSLTrustStore=FILE;SSLTrustStorePwd=PASS Facing below…
0
votes
0 answers

how to access the NiFi and Sqoop 1 services?

I am writing to inquire about my 60-day CDH7 trial. I have been trying to create a cluster with NiFi and Sqoop 1 services, but I have been unable to locate them. After conducting some research, I discovered that I need the CFM parcel to access these…
0
votes
0 answers

CDH(Cloudera Distributed Hadoop) to CDP(Cloudera Data Platform) migration Spark 1x-3x query

We are currently doing a feasibility study on migrating from CDH(Cloudera Distributed Hadoop) to CDP(Cloudera Data Platform) wrt spark(currently in version 1.6). When checked the documenation,it was understood that 1.6 is not supported ,we need to…
0
votes
1 answer

Issue of container OOM when writing Dataframe to parquet files in Spark Job

I'm using Machine Learning Workspace in Cloudera Data Platform (CDP). I created a session with 4vCPU/16 GiB Memory and enabled Spark 3.2.0. I'm using spark to load data of one month (the whole month data size is around 12 GB) and do some…
Ryan
  • 33
  • 4
0
votes
1 answer

Connection to remote Hadoop Cluster (CDP) through Linux server

I'm new to PySpark and I want to connect remote Hadoop Cluster (CDP) through Linux server by using spark-submit command. Any help would be appreciated. I need spark-submit command to connect remote CDP.
1
2