You can translate the question and the replies:

querying very large data sets in denodo?

if I want to build a machine learning model using Apache Spark that accesses terabytes of data in denodo, is there a way to access the data using a more performant method than jdbc, such as hdfs?
user
07-01-2020 07:32:24 -0500

1 Answer

Hi, For connections to HDFS data sources, the Denodo Distributed File System Custom Wrapper is provided on the [Denodo Support Site](https://support.denodo.com/) for connections to HDFS, Amazon S3, Azure Blob Storage, Azure Data Lake Storage and Google Cloud Storage. The Denodo Platform also includes built in methods for connecting to certain Hadoop data sources; supported sources and connection methods are detailed in the [Denodo and Hadoop](https://community.denodo.com/kb/view/document/Denodo%20and%20Hadoop?category=Data+Sources) page of the Knowledge Base. I’ve included the link to the [Distributed File System Custom Wrapper user manual](https://community.denodo.com/docs/html/document/denodoconnects/7.0/Denodo%20Distributed%20File%20System%20Custom%20Wrapper%20-%20User%20Manual) here. Hope this helps!
Denodo Team
09-01-2020 06:06:21 -0500
You must sign in to add an answer. If you do not have an account, you can register here