Best Practices for connecting to HDFS 3.x and Hive 2.1.1 - spectrum_platform - 23.1

spectrum-inline

Product type
Software
Portfolio
Enrich
Integrate
Locate
Verify
Product family
Spectrum
Product
Spectrum > GeoEnrichment
Spectrum > Spectrum Platform
Spectrum > Dataflow Designer
Spectrum > Quality > Spectrum Quality
Spectrum > Discovery
Spectrum > Quality > Context Graph
Spectrum > Quality > Addressing
Spectrum > Web Services
Spectrum > Data Integration
Spectrum > Spatial > Spectrum Spatial
Spectrum > Geo Addressing > Enterprise Tax
Version
23.1
Language
English
Product name
Precisely Spectrum
Title
spectrum-inline
First publish date
2007
Last updated
2023-06-02
Published on
2023-06-02T09:54:39.526000
  • While creating tables and defining schema, use lower casing. Example: Create Table demo ( id int, name string, salary int).
  • Use fully qualified field names when using Hive JDBC connection with Read from DB stage. Else, for reading from DB and writing to it, create a Model Store connection.
  • Location for Keytab file to specify forwarded slashes / when connecting from Windows.
  • Avoid spaces in Hive file, table, and field names.
  • Use the property grid for creating the HDFS connection
  • Creating HDFS connection
    • If your cluster is secured, or in other words, it is HA Enabled or SSL Enabled, ensure this: The value of the property = fs.defaultFS is swebhdfs://<Nameservice name>, where swebhdfs stands for secured webhdfs.
    • If cluster is not secured, you can use webhdfs. Besides, nameservice needs to be provided as it is used for other properties.
  • This entry needs to be made in the Wrapper.conf file: wrapper.java.additional.7=-Djavax.net.ssl.trustStore=<Path of truststore file>/<name of the truststore file>