Skip to content

Commit

Permalink
add hdfs error for dag (#1761)
Browse files Browse the repository at this point in the history
  • Loading branch information
cooper-lzy authored Nov 17, 2022
1 parent 793d2ce commit 00032be
Show file tree
Hide file tree
Showing 3 changed files with 22 additions and 1 deletion.
6 changes: 6 additions & 0 deletions docs-2.0/nebula-explorer/deploy-connect/ex-ug-deploy.md
Original file line number Diff line number Diff line change
Expand Up @@ -29,6 +29,12 @@ Before deploying Explorer, you must check the following information:

License is only available in the Enterprise Edition. To obtain the license, apply for [NebulaGraph Explorer Free Trial](https://nebula-graph.io/visualization-tools-free-trial).

- The HDFS services are deployed if graph computing is required. The namenode uses port 8020 by default, and the datanode uses port 50010 by default.

!!! caution

If the HDFS port is unavailable, the connection timeout message may be displayed.

## RPM-based deployment

### Installation
Expand Down
13 changes: 13 additions & 0 deletions docs-2.0/nebula-explorer/faq.md
Original file line number Diff line number Diff line change
Expand Up @@ -68,6 +68,19 @@ Check according to the following procedure:

3. Restart the Dag Controller for the settings to take effect.

## How to resolve the error `no available namenodes: dial tcp xx.xx.xx.xx:8020: connect: connection timed out`?

Check whether the HDFS namenode port 8020 is open.

## How to resolve the error `org.apache.hadoop.net.ConnectTimeoutException: 60000 millis timeout`?

Check whether the HDFS datanode port 50010 is open.

If the port is not opened, an error similar to the following may be reported:

- `Check failed: false close hdfs-file failed`
- `org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /analytics/xx/tasks/analytics_xxx/xxx.csv could only be replicated to 0 nodes instead of minReplication`

## How to resolve the error `broadcast.hpp:193] Check failed: (size_t)recv_bytes >= sizeof(chunk_tail_t) recv message too small: 0`?

The amount of data to be processed is too small, but the number of compute nodes and processes is too large. Smaller `clusterSize` and `processes` need to be set when submitting jobs.
4 changes: 3 additions & 1 deletion docs-2.0/nebula-explorer/workflow/1.prepare-resources.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,10 +16,12 @@ You must prepare your environment for running a workflow, including NebulaGraph

3. Configure the following resources:

![workflow_configuration](https://docs-cdn.nebula-graph.com.cn/figures/workflow_configuration_221117_en.png)

|Type|Description|
|:--|:--|
|NebulaGraph Configuration| The access address of the graph service that executes a graph query or to which the graph computing result is written. The default address is the address that you use to log into Explorer and can not be changed. You can set timeout periods for three services. |
|HDFS Configuration| The HDFS address that stores the result of the graph query or graph computing (`fs.defaultFS`). Click **Add** to add a new address, you can set the HDFS name, HDFS path, and HDFS username (optional). The configuration takes effect only after the HDFS client is installed on the machine where the Analytics is installed. |
|HDFS Configuration| The HDFS address that stores the result of the graph query or graph computing. Click **Add** to add a new address, you can set the HDFS name, HDFS path (`fs.defaultFS`), and HDFS username. You can configure the save path, such as `hdfs://192.168.8.100:9000/test`. The configuration takes effect only after the HDFS client is installed on the machine where the Analytics is installed. |
|NebulaGraph Analytics Configuration| The NebulaGraph Analytics address that performs the graph computing. Click **Add** to add a new address.|

4. Click **Confirm**.

0 comments on commit 00032be

Please sign in to comment.