Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

HiveWarehouseConnector fetching only 1000 rows through hive.execute() #279

Open
shesadri opened this issue Dec 3, 2019 · 2 comments
Open

Comments

@shesadri
Copy link

shesadri commented Dec 3, 2019

we are using the spark 2.3.0 version with Hadoop3 to fetch the records using hive table. While using the hive connector library we are facing issue where it does only 1000 records fetch though we have more than millions records eligible for query we pass. Any possibility of overriding this value set for us to get more records.

@justinleet
Copy link

I saw this issue a bit ago. execute looks like it only runs through the driver (and increasing that limit can OOM) and should be used primarily for catalog operations. The solution was to use executeQuery instead.

@guerinclement
Copy link

I saw this issue a bit ago. execute looks like it only runs through the driver (and increasing that limit can OOM) and should be used primarily for catalog operations. The solution was to use executeQuery instead.

Yes, but the LIMIT instruction used in executeQuery() does not return the required number of lines (up to 260x from my tests).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants