Best Practices for connecting to HDFS 3.x and Hive 2.1.1 - spectrum_platform - 23.1
spectrum-inline
- Product type
- Software
- Portfolio
- Enrich
- Integrate
- Locate
- Verify
- Product family
- Spectrum
- Product
- Spectrum > GeoEnrichment
- Spectrum > Spectrum Platform
- Spectrum > Dataflow Designer
- Spectrum > Quality > Spectrum Quality
- Spectrum > Discovery
- Spectrum > Quality > Context Graph
- Spectrum > Quality > Addressing
- Spectrum > Web Services
- Spectrum > Data Integration
- Spectrum > Spatial > Spectrum Spatial
- Spectrum > Geo Addressing > Enterprise Tax
- Version
- 23.1
- Language
- English
- Product name
- Precisely Spectrum
- Title
- spectrum-inline
- First publish date
- 2007
- Last updated
- 2023-06-02
- Published on
- 2023-06-02T09:54:39.526000
- While creating tables and defining schema, use lower casing. Example: Create Table demo ( id
int, name string, salary int).
- Use fully qualified field names when using Hive JDBC connection with Read from
DB stage. Else, for reading from DB and writing to it, create a Model
Store connection.
- Location for Keytab file to specify forwarded slashes / when
connecting from Windows.
- Avoid spaces in Hive file, table, and field names.
- Use the property grid for creating the HDFS connection
- Creating HDFS connection
- If your cluster is secured, or in other words, it is HA Enabled or SSL Enabled, ensure
this: The value of the property = fs.defaultFS is
swebhdfs://<Nameservice name>, where swebhdfs stands
for secured webhdfs.
- If cluster is not secured, you can use webhdfs. Besides,
nameservice needs to be provided as it is used for other properties.
- This entry needs to be made in the Wrapper.conf file:
wrapper.java.additional.7=-Djavax.net.ssl.trustStore=<Path of truststore
file>/<name of the truststore file>