Jump to content

Client daggara Spark Sahayam kavali.... urgent


powerking

Recommended Posts

Hi, 

Urgent ga Spark sahayam kavali client daggara unna.... stuck with some serious issues... help or refer me anyone who does this kind of stuff..

will Pay Too....

Link to comment
Share on other sites

1 hour ago, powerking said:

Hi, 

Urgent ga Spark sahayam kavali client daggara unna.... stuck with some serious issues... help or refer me anyone who does this kind of stuff..

will Pay Too....

Spark lo em use chestunav which tool/language 

linkedin itversity group lo Chudu man help dorkachu

ekda stuck ayyav aa query ikda veyyi it would be helpful

Link to comment
Share on other sites

9 minutes ago, vendetta said:

Spark lo em use chestunav which tool/language 

linkedin itversity group lo Chudu man help dorkachu

ekda stuck ayyav aa query ikda veyyi it would be helpful

Using PySpark ... trying to connect Hive tables and working on querying to get result sets for analysis.... 

My query is, how do we create Hive tables with 100 columns on files which are in hdfs... later I want to query these tables in spark ( preferably Pyspark) for analytics.... 

I know I can do this in Spark SQL but Indexing is something I wanna consider here.... correct me 

I'll be heading to Linkedin anyway.... Thanks for the response 

Please do let me know if anyone worked on the scenario mentioned above...

Peace Out !

Link to comment
Share on other sites

1 hour ago, powerking said:

Using PySpark ... trying to connect Hive tables and working on querying to get result sets for analysis.... 

My query is, how do we create Hive tables with 100 columns on files which are in hdfs... later I want to query these tables in spark ( preferably Pyspark) for analytics.... 

I know I can do this in Spark SQL but Indexing is something I wanna consider here.... correct me 

I'll be heading to Linkedin anyway.... Thanks for the response 

Please do let me know if anyone worked on the scenario mentioned above...

Peace Out !

first create table in Hive (Managed or External ---depends on ur requirement):

Sample code:

create table tab_name(

column1 int,

column2 string....)

row format delimited fields terminated by '\t';

use the above sample code and create a table for all 100 columns.

Then,  load data from hdfs to hive by below code:

load data inpath "hdfs:///user/....." into table tab_name;

-------------------------------------------------------------------------------------

Pyspark analytics ki sample data cheppu...I will try to write query

 

Link to comment
Share on other sites

Note: If you use Managed table like I said in above sample code, the hdfs data will be moved from hdfs to hive. so there won't be any data in hdfs after u move to hive. Anyways check according to your requirements. I would suggest you to use external table like below. 

Sample code:

create external table tab_name(

column1 int,

column2 string....)

row format delimited fields terminated by '\t'

LOCATION /user/...... ;

 

 

 

Link to comment
Share on other sites

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

×
×
  • Create New...