MongoDb
Description
Read data from MongoDB.
Engine Supported and plugin name
- Spark: MongoDb
- Flink
Options
| name | type | required | default value |
|---|---|---|---|
| readconfig.uri | string | yes | - |
| readconfig.database | string | yes | - |
| readconfig.collection | string | yes | - |
| readconfig.* | string | no | - |
| schema | string | no | - |
| common-options | string | yes | - |
readconfig.uri [string]
MongoDB uri
readconfig.database [string]
MongoDB database
readconfig.collection [string]
MongoDB collection
readconfig.* [string]
More other parameters can be configured here, see MongoDB Configuration for details, see the Input Configuration section. The way to specify parameters is to prefix the original parameter name readconfig. For example, the way to set spark.mongodb.input.partitioner is readconfig.spark.mongodb.input.partitioner="MongoPaginateBySizePartitioner" . If you do not specify these optional parameters, the default values of the official MongoDB documentation will be used.
schema [string]
Because MongoDB does not have the concept of schema, when spark reads MongoDB , it will sample MongoDB data and infer the schema . In fact, this process will be slow and may be inaccurate. This parameter can be manually specified. Avoid these problems. schema is a json string, such as {\"name\":\"string\",\"age\":\"integer\",\"addrs\":{\"country\":\"string\ ",\"city\":\"string\"}}
common options [string]
Source Plugin common parameters, refer to Source Plugin for details
Example
mongodb {
readconfig.uri = "mongodb://username:password@127.0.0.1:27017/mypost"
readconfig.database = "mydatabase"
readconfig.collection = "mycollection"
readconfig.spark.mongodb.input.partitioner = "MongoPaginateBySizePartitioner"
schema="{\"name\":\"string\",\"age\":\"integer\",\"addrs\":{\"country\":\"string\",\"city\":\"string\"}}"
result_table_name = "mongodb_result_table"
}