My Hours
My Hours source connector
Support Those Engines
Spark
Flink
SeaTunnel Zeta
Key Features
Description
Used to read data from My Hours.
Key features
Supported DataSource Info
In order to use the My Hours connector, the following dependencies are required. They can be downloaded via install-plugin.sh or from the Maven central repository.
Datasource | Supported Versions | Dependency |
---|---|---|
My Hours | universal | Download |
Source Options
Name | Type | Required | Default | Description |
---|---|---|---|---|
url | String | Yes | - | Http request url. |
String | Yes | - | My hours login email address. | |
password | String | Yes | - | My hours login password. |
schema | Config | No | - | Http and seatunnel data structure mapping |
schema.fields | Config | No | - | The schema fields of upstream data |
json_field | Config | No | - | This parameter helps you configure the schema,so this parameter must be used with schema. |
content_json | String | No | - | This parameter can get some json data.If you only need the data in the 'book' section, configure content_field = "$.store.book.*" . |
format | String | No | json | The format of upstream data, now only support json text , default json . |
method | String | No | get | Http request method, only supports GET, POST method. |
headers | Map | No | - | Http headers. |
params | Map | No | - | Http params. |
body | String | No | - | Http body. |
poll_interval_ms | Int | No | - | Request http api interval(millis) in stream mode. |
retry | Int | No | - | The max retry times if request http return to IOException . |
retry_backoff_multiplier_ms | Int | No | 100 | The retry-backoff times(millis) multiplier if request http failed. |
retry_backoff_max_ms | Int | No | 10000 | The maximum retry-backoff times(millis) if request http failed |
enable_multi_lines | Boolean | No | false | |
common-options | No | - | Source plugin common parameters, please refer to Source Common Options for details |
How to Create a My Hours Data Synchronization Jobs
env {
execution.parallelism = 1
job.mode = "BATCH"
}
MyHours{
url = "https://api2.myhours.com/api/Projects/getAll"
email = "seatunnel@test.com"
password = "seatunnel"
schema {
fields {
name = string
archived = boolean
dateArchived = string
dateCreated = string
clientName = string
budgetAlertPercent = string
budgetType = int
totalTimeLogged = double
budgetValue = double
totalAmount = double
totalExpense = double
laborCost = double
totalCost = double
billableTimeLogged = double
totalBillableAmount = double
billable = boolean
roundType = int
roundInterval = int
budgetSpentPercentage = double
budgetTarget = int
budgetPeriodType = string
budgetSpent = string
id = string
}
}
}
# Console printing of the read data
sink {
Console {
parallelism = 1
}
}
Parameter Interpretation
format
when you assign format is json
, you should also assign schema option, for example:
upstream data is the following:
{
"code": 200,
"data": "get success",
"success": true
}
you should assign schema as the following:
schema {
fields {
code = int
data = string
success = boolean
}
}
connector will generate data as the following:
code | data | success |
---|---|---|
200 | get success | true |
when you assign format is text
, connector will do nothing for upstream data, for example:
upstream data is the following:
{
"code": 200,
"data": "get success",
"success": true
}
connector will generate data as the following:
content |
---|
{"code": 200, "data": "get success", "success": true} |
content_json
This parameter can get some json data.If you only need the data in the 'book' section, configure content_field = "$.store.book.*"
.
If your return data looks something like this.
{
"store": {
"book": [
{
"category": "reference",
"author": "Nigel Rees",
"title": "Sayings of the Century",
"price": 8.95
},
{
"category": "fiction",
"author": "Evelyn Waugh",
"title": "Sword of Honour",
"price": 12.99
}
],
"bicycle": {
"color": "red",
"price": 19.95
}
},
"expensive": 10
}
You can configure content_field = "$.store.book.*"
and the result returned looks like this:
[
{
"category": "reference",
"author": "Nigel Rees",
"title": "Sayings of the Century",
"price": 8.95
},
{
"category": "fiction",
"author": "Evelyn Waugh",
"title": "Sword of Honour",
"price": 12.99
}
]
Then you can get the desired result with a simpler schema,like
Http {
url = "http://mockserver:1080/contentjson/mock"
method = "GET"
format = "json"
content_field = "$.store.book.*"
schema = {
fields {
category = string
author = string
title = string
price = string
}
}
}
Here is an example:
- Test data can be found at this link mockserver-config.json
- See this link for task configuration http_contentjson_to_assert.conf.
json_field
This parameter helps you configure the schema,so this parameter must be used with schema.
If your data looks something like this:
{
"store": {
"book": [
{
"category": "reference",
"author": "Nigel Rees",
"title": "Sayings of the Century",
"price": 8.95
},
{
"category": "fiction",
"author": "Evelyn Waugh",
"title": "Sword of Honour",
"price": 12.99
}
],
"bicycle": {
"color": "red",
"price": 19.95
}
},
"expensive": 10
}
You can get the contents of 'book' by configuring the task as follows:
source {
Http {
url = "http://mockserver:1080/jsonpath/mock"
method = "GET"
format = "json"
json_field = {
category = "$.store.book[*].category"
author = "$.store.book[*].author"
title = "$.store.book[*].title"
price = "$.store.book[*].price"
}
schema = {
fields {
category = string
author = string
title = string
price = string
}
}
}
}
- Test data can be found at this link mockserver-config.json
- See this link for task configuration http_jsonpath_to_assert.conf.
Changelog
next version
- Add My Hours Source Connector
- [Feature][Connector-V2][HTTP] Use json-path parsing (3510)