...
Limitations and Known Issues
None.
Snap Views
Type | Format | Number of Views | Examples of Upstream and Downstream Snaps | Description |
---|---|---|---|---|
Input | Document |
|
| The file filter, file, and directory details of the file to be deleted. |
Output | Document |
|
| The deleted file or a group of files. |
Error | Error handling is a generic way to handle errors without losing data or failing the Snap execution. You can handle the errors that the Snap might encounter when running the Pipeline by choosing one of the following options from the When errors occur list under the Views tab:
Learn more about Error handling in Pipelines. |
Snap Settings
Info |
---|
|
Field Name | Field Type | Description | |||||||
---|---|---|---|---|---|---|---|---|---|
Label*
Default Value: HDFS delete | String | The name for the Snap. You can modify this to be more specific, especially if you have more than one of the same Snap in your pipeline. | |||||||
Directory
Default Value: hdfs://<hostname>:<port>/
| String/Expression/Suggestion | Specify the URL for the HDFS directory. It should start with the HDFS file protocol in the following format:
The Directory property is used only in the Suggest operation. When you click the Suggestion icon, the Snap displays a list of subdirectories under the specific directory. It generates the list by applying the value specified in the File Filter property. | |||||||
File filter
Default Value: * Example: ?
| String/Expression | Specify the Glob filter pattern. A file filter is a criteria to include or exclude specific files when processing data in HDFS.
| |||||||
File
Default Value: N/A
| String/Expression/Suggestion | Specify the file name or a relative path to a file under the directory specified in the Directory property. It should not start with a URL separator "/". The value of the File property depends on the name of the directory specified in the Directory property and the criterion specified in the File filter property. | |||||||
User Impersonation
Default Value: Deselected | Checkbox | Select this checkbox to enable user impersonation. Hadoop allows you to configure proxy users to access HDFS on behalf of other users; this is called impersonation. When user impersonation is enabled on the Hadoop cluster, any jobs submitted using a proxy are executed with the impersonated user's existing privilege levels rather than those of the superuser associated with the cluster. For more information on user impersonation in this Snap, refer to the section on User Impersonation below. | |||||||
Delete Directory Default Value: Deselected | Checkbox/Expression | Select this checkbox to enable you to delete all the paths in the specified directory. | |||||||
Number Of Retries
Default Value: 0 Example: 12 | Integer/Expression | Specify the maximum number of attempts to be made to receive a response.
| |||||||
Retry Interval (seconds)
Default Value: 1 Example: 30 | Integer/Expression | Specify the time interval between two successive retry requests. A retry happens only when the previous attempt resulted in an exception. | |||||||
Snap Execution Default Value: Execute Only | Dropdown list | Select one of the following three modes in which the Snap executes:
|
Troubleshooting
Error | Reason | Resolution |
---|---|---|
Remote filesystem access failed. | The user credentials or URL might be incorrect, or the remote server may be inaccessible. It indicates a problem with the communication between the nodes in your Hadoop cluster or an issue with the underlying HDFS. | Check the user credentials and URL and retry. Check the permissions and access rights of the Hadoop files and directories. Ensure that you have the required permissions to access and modify the data. |
A directory is not a valid string. | The expression or value specified in the Directory property is either not existing in HDFS or not accessible. | Please check if a valid expression is entered in the Directory property and if the correct document data is at the input view. |
Deleting multiple JSON files from Azure Data Lake Storage
...
Configure the HDFS Writer Snap with specific details, such as the destination directory where the files should be added in the Azure Data Lake Storage. You can see that the file is written to the Azure Data Lake Storage in the output preview.
Snap configuration | Output preview |
---|---|
You can delete the same file from the Azure Data Lake Storage with the HDFS delete Snap.
Snap Configuration | Output preview |
---|---|
Downloads
Info |
---|
|
...
...