2021-02-02 07:14:38 +00:00
|
|
|
import boto3
|
|
|
|
import botocore.credentials
|
|
|
|
from botocore.awsrequest import AWSRequest
|
|
|
|
from botocore.endpoint import URLLib3Session
|
|
|
|
from botocore.auth import SigV4Auth
|
|
|
|
import json
|
|
|
|
import os
|
2021-02-22 06:13:30 +00:00
|
|
|
from datetime import datetime, timedelta, timezone
|
|
|
|
import sys, traceback
|
2021-04-20 08:06:21 +00:00
|
|
|
import re
|
2021-04-27 22:30:28 +00:00
|
|
|
import html
|
2021-07-18 22:38:31 +00:00
|
|
|
import base64
|
|
|
|
import gzip
|
|
|
|
from io import BytesIO
|
|
|
|
|
2021-11-29 10:09:32 +00:00
|
|
|
from multiprocessing import Process
|
|
|
|
|
|
|
|
http_session = URLLib3Session()
|
|
|
|
|
|
|
|
def mirror(path,params):
|
|
|
|
session = boto3.Session()
|
|
|
|
headers = {"Host": "search-sondes-v2-hiwdpmnjbuckpbwfhhx65mweee.us-east-1.es.amazonaws.com", "Content-Type": "application/json", "Content-Encoding":"gzip"}
|
|
|
|
request = AWSRequest(
|
|
|
|
method="POST", url=f"https://search-sondes-v2-hiwdpmnjbuckpbwfhhx65mweee.us-east-1.es.amazonaws.com/{path}", data=params, headers=headers
|
|
|
|
)
|
|
|
|
SigV4Auth(boto3.Session().get_credentials(), "es", "us-east-1").add_auth(request)
|
|
|
|
session = URLLib3Session()
|
|
|
|
r = session.send(request.prepare())
|
2021-02-02 07:14:38 +00:00
|
|
|
|
2021-02-02 10:44:39 +00:00
|
|
|
HOST = os.getenv("ES")
|
2021-02-02 07:14:38 +00:00
|
|
|
# get current sondes, filter by date, location
|
|
|
|
|
2021-02-02 10:44:39 +00:00
|
|
|
|
2021-02-02 07:14:38 +00:00
|
|
|
def get_sondes(event, context):
|
|
|
|
path = "telm-*/_search"
|
|
|
|
payload = {
|
|
|
|
"aggs": {
|
|
|
|
"2": {
|
2021-02-02 10:44:39 +00:00
|
|
|
"terms": {
|
|
|
|
"field": "serial.keyword",
|
|
|
|
"order": {"_key": "desc"},
|
|
|
|
"size": 10000,
|
2021-02-02 07:14:38 +00:00
|
|
|
},
|
2021-02-02 10:44:39 +00:00
|
|
|
"aggs": {
|
|
|
|
"1": {
|
|
|
|
"top_hits": {
|
|
|
|
"size": 1,
|
|
|
|
"sort": [{"datetime": {"order": "desc"}}],
|
2021-02-02 07:14:38 +00:00
|
|
|
}
|
|
|
|
}
|
2021-02-02 10:44:39 +00:00
|
|
|
},
|
2021-02-02 07:14:38 +00:00
|
|
|
}
|
|
|
|
},
|
2021-02-02 10:44:39 +00:00
|
|
|
"query": {"bool": {"filter": [{"match_all": {}}]}},
|
2021-02-02 07:14:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
# add filters
|
|
|
|
if "queryStringParameters" in event:
|
|
|
|
if "last" in event["queryStringParameters"]:
|
|
|
|
payload["query"]["bool"]["filter"].append(
|
|
|
|
{
|
|
|
|
"range": {
|
|
|
|
"datetime": {
|
|
|
|
"gte": f"now-{int(event['queryStringParameters']['last'])}s",
|
2021-07-24 09:33:29 +00:00
|
|
|
"lte": "now+1m",
|
2021-02-02 07:14:38 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
)
|
2021-08-20 02:49:55 +00:00
|
|
|
else:
|
|
|
|
payload["query"]["bool"]["filter"].append(
|
|
|
|
{"range": {"datetime": {"gte": "now-1d", "lte": "now+1m"}}}
|
|
|
|
)
|
|
|
|
|
2021-02-02 10:44:39 +00:00
|
|
|
if (
|
|
|
|
"lat" in event["queryStringParameters"]
|
|
|
|
and "lon" in event["queryStringParameters"]
|
|
|
|
and "distance" in event["queryStringParameters"]
|
|
|
|
):
|
2021-02-02 07:14:38 +00:00
|
|
|
payload["query"]["bool"]["filter"].append(
|
|
|
|
{
|
|
|
|
"geo_distance": {
|
|
|
|
"distance": f"{int(event['queryStringParameters']['distance'])}m",
|
|
|
|
"position": {
|
2021-02-02 10:44:39 +00:00
|
|
|
"lat": float(event["queryStringParameters"]["lat"]),
|
|
|
|
"lon": float(event["queryStringParameters"]["lon"]),
|
|
|
|
},
|
2021-02-02 07:14:38 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
)
|
2021-08-20 02:49:55 +00:00
|
|
|
else:
|
2021-02-02 07:14:38 +00:00
|
|
|
payload["query"]["bool"]["filter"].append(
|
2021-08-20 02:49:55 +00:00
|
|
|
{"range": {"datetime": {"gte": "now-1d", "lte": "now+1m"}}}
|
2021-02-02 10:44:39 +00:00
|
|
|
)
|
2021-08-20 02:49:55 +00:00
|
|
|
|
2021-02-02 10:44:39 +00:00
|
|
|
results = es_request(payload, path, "POST")
|
|
|
|
buckets = results["aggregations"]["2"]["buckets"]
|
|
|
|
sondes = {
|
|
|
|
bucket["1"]["hits"]["hits"][0]["_source"]["serial"]: bucket["1"]["hits"][
|
|
|
|
"hits"
|
|
|
|
][0]["_source"]
|
|
|
|
for bucket in buckets
|
|
|
|
}
|
|
|
|
return json.dumps(sondes)
|
|
|
|
|
|
|
|
|
|
|
|
def get_telem(event, context):
|
|
|
|
|
|
|
|
durations = { # ideally we shouldn't need to predefine these, but it's a shit load of data and we don't need want to overload ES
|
|
|
|
"3d": (259200, 1200), # 3d, 20m
|
|
|
|
"1d": (86400, 600), # 1d, 10m
|
2021-07-19 05:54:54 +00:00
|
|
|
"12h": (43200, 600), # 1d, 10m
|
2021-09-17 13:50:36 +00:00
|
|
|
"6h": (21600, 120), # 6h, 1m
|
|
|
|
"3h": (10800, 60), # 3h, 10s
|
|
|
|
"1h": (3600, 40),
|
|
|
|
"30m": (1800, 20),
|
2021-07-19 06:26:50 +00:00
|
|
|
"1m": (60, 1),
|
2021-07-26 06:58:47 +00:00
|
|
|
"15s": (15, 1),
|
|
|
|
"0": (0, 1) # for getting a single time point
|
2021-02-02 10:44:39 +00:00
|
|
|
}
|
|
|
|
duration_query = "3h"
|
2021-04-09 07:14:14 +00:00
|
|
|
requested_time = datetime.now(timezone.utc)
|
2021-02-15 05:23:51 +00:00
|
|
|
|
2021-02-02 10:44:39 +00:00
|
|
|
if (
|
|
|
|
"queryStringParameters" in event
|
|
|
|
and "duration" in event["queryStringParameters"]
|
|
|
|
):
|
|
|
|
if event["queryStringParameters"]["duration"] in durations:
|
|
|
|
duration_query = event["queryStringParameters"]["duration"]
|
|
|
|
else:
|
|
|
|
return f"Duration must be either {', '.join(durations.keys())}"
|
|
|
|
|
2021-02-15 05:23:51 +00:00
|
|
|
if (
|
|
|
|
"queryStringParameters" in event
|
|
|
|
and "datetime" in event["queryStringParameters"]
|
|
|
|
):
|
2021-03-28 05:00:44 +00:00
|
|
|
requested_time = datetime.fromisoformat(
|
|
|
|
event["queryStringParameters"]["datetime"].replace("Z", "+00:00")
|
|
|
|
)
|
2021-02-15 05:23:51 +00:00
|
|
|
|
2021-02-02 10:44:39 +00:00
|
|
|
(duration, interval) = durations[duration_query]
|
2021-07-19 06:26:50 +00:00
|
|
|
if "serial" in event["queryStringParameters"]:
|
|
|
|
interval = 1
|
2021-07-26 06:58:47 +00:00
|
|
|
lt = requested_time + timedelta(0, 1)
|
2021-03-28 05:00:44 +00:00
|
|
|
gte = requested_time - timedelta(0, duration)
|
2021-02-02 10:44:39 +00:00
|
|
|
|
2021-08-01 13:47:51 +00:00
|
|
|
path = f"telm-{lt.year:2}-{lt.month:02},telm-{gte.year:2}-{gte.month:02}/_search"
|
2021-02-02 10:44:39 +00:00
|
|
|
payload = {
|
2021-07-19 01:10:14 +00:00
|
|
|
"timeout": "30s",
|
2021-08-01 13:47:51 +00:00
|
|
|
"size": 0,
|
2021-02-02 10:44:39 +00:00
|
|
|
"aggs": {
|
|
|
|
"2": {
|
|
|
|
"terms": {
|
|
|
|
"field": "serial.keyword",
|
|
|
|
"order": {"_key": "desc"},
|
|
|
|
"size": 10000,
|
|
|
|
},
|
|
|
|
"aggs": {
|
|
|
|
"3": {
|
|
|
|
"date_histogram": {
|
|
|
|
"field": "datetime",
|
|
|
|
"fixed_interval": f"{str(interval)}s",
|
|
|
|
"min_doc_count": 1,
|
|
|
|
},
|
|
|
|
"aggs": {
|
|
|
|
"1": {
|
|
|
|
"top_hits": {
|
2021-02-02 11:09:04 +00:00
|
|
|
# "docvalue_fields": [
|
|
|
|
# {"field": "position"},
|
|
|
|
# {"field": "alt"},
|
|
|
|
# {"field": "datetime"},
|
|
|
|
# ],
|
|
|
|
# "_source": "position",
|
2021-08-01 13:47:51 +00:00
|
|
|
"size": 10 if (duration == 0 ) else 1,
|
2021-07-26 09:51:48 +00:00
|
|
|
"sort": [
|
|
|
|
{"datetime": {"order": "desc"}},
|
2021-08-20 13:41:29 +00:00
|
|
|
{"pressure": {"order": "desc","mode" : "median"}}
|
2021-07-26 09:51:48 +00:00
|
|
|
],
|
2021-02-02 10:44:39 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
}
|
|
|
|
},
|
|
|
|
}
|
|
|
|
},
|
|
|
|
"query": {
|
|
|
|
"bool": {
|
2021-07-21 09:54:09 +00:00
|
|
|
"must_not": [{"match_phrase": {"software_name": "SondehubV1"}}, {"match_phrase": {"serial": "xxxxxxxx"}}],
|
2021-02-02 10:44:39 +00:00
|
|
|
"filter": [
|
|
|
|
{"match_all": {}},
|
|
|
|
{
|
|
|
|
"range": {
|
2021-03-28 05:00:44 +00:00
|
|
|
"datetime": {"gte": gte.isoformat(), "lt": lt.isoformat()}
|
2021-02-02 07:14:38 +00:00
|
|
|
}
|
2021-02-02 11:09:04 +00:00
|
|
|
},
|
2021-02-02 10:44:39 +00:00
|
|
|
]
|
|
|
|
}
|
|
|
|
},
|
|
|
|
}
|
|
|
|
if "queryStringParameters" in event:
|
|
|
|
if "serial" in event["queryStringParameters"]:
|
|
|
|
payload["query"]["bool"]["filter"].append(
|
|
|
|
{
|
|
|
|
"match_phrase": {
|
|
|
|
"serial": str(event["queryStringParameters"]["serial"])
|
|
|
|
}
|
2021-02-02 07:14:38 +00:00
|
|
|
}
|
|
|
|
)
|
|
|
|
results = es_request(payload, path, "POST")
|
2021-02-02 10:44:39 +00:00
|
|
|
output = {
|
|
|
|
sonde["key"]: {
|
2021-07-26 09:51:48 +00:00
|
|
|
data["key_as_string"]: dict(data["1"]["hits"]["hits"][0]["_source"],
|
|
|
|
uploaders=[ #add additional uploader information
|
2021-10-27 00:00:05 +00:00
|
|
|
{key:value for key,value in uploader['_source'].items() if key in ["snr","rssi","uploader_callsign", "frequency"]}
|
2021-07-26 09:51:48 +00:00
|
|
|
for uploader in data["1"]["hits"]["hits"]
|
|
|
|
])
|
2021-02-02 10:44:39 +00:00
|
|
|
for data in sonde["3"]["buckets"]
|
|
|
|
}
|
|
|
|
for sonde in results["aggregations"]["2"]["buckets"]
|
|
|
|
}
|
2021-03-28 05:00:44 +00:00
|
|
|
|
2021-07-21 09:54:09 +00:00
|
|
|
compressed = BytesIO()
|
|
|
|
with gzip.GzipFile(fileobj=compressed, mode='w') as f:
|
|
|
|
json_response = json.dumps(output)
|
|
|
|
f.write(json_response.encode('utf-8'))
|
|
|
|
|
|
|
|
gzippedResponse = compressed.getvalue()
|
|
|
|
return {
|
|
|
|
"body": base64.b64encode(gzippedResponse).decode(),
|
|
|
|
"isBase64Encoded": True,
|
|
|
|
"statusCode": 200,
|
|
|
|
"headers": {
|
|
|
|
"Content-Encoding": "gzip",
|
|
|
|
"content-type": "application/json"
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
def get_listener_telemetry(event, context):
|
|
|
|
|
|
|
|
durations = { # ideally we shouldn't need to predefine these, but it's a shit load of data and we don't need want to overload ES
|
2021-08-20 13:41:29 +00:00
|
|
|
"3d": (259200, 2400), # 3d, 20m
|
|
|
|
"1d": (86400, 2400), # 1d, 10m
|
|
|
|
"12h": (43200, 1200), # 1d, 10m
|
|
|
|
"6h": (21600, 300), # 6h, 1m
|
|
|
|
"3h": (10800, 120), # 3h, 10s
|
|
|
|
"1h": (3600, 120),
|
|
|
|
"30m": (1800, 30),
|
2021-07-21 09:54:09 +00:00
|
|
|
"1m": (60, 1),
|
2021-08-20 13:41:29 +00:00
|
|
|
"15s": (15, 1),
|
|
|
|
"0": (0, 1)
|
2021-07-21 09:54:09 +00:00
|
|
|
}
|
|
|
|
duration_query = "3h"
|
|
|
|
requested_time = datetime.now(timezone.utc)
|
|
|
|
|
|
|
|
if (
|
|
|
|
"queryStringParameters" in event
|
|
|
|
and "duration" in event["queryStringParameters"]
|
|
|
|
):
|
|
|
|
if event["queryStringParameters"]["duration"] in durations:
|
|
|
|
duration_query = event["queryStringParameters"]["duration"]
|
|
|
|
else:
|
|
|
|
return f"Duration must be either {', '.join(durations.keys())}"
|
|
|
|
|
|
|
|
if (
|
|
|
|
"queryStringParameters" in event
|
|
|
|
and "datetime" in event["queryStringParameters"]
|
|
|
|
):
|
|
|
|
requested_time = datetime.fromisoformat(
|
|
|
|
event["queryStringParameters"]["datetime"].replace("Z", "+00:00")
|
|
|
|
)
|
|
|
|
|
|
|
|
(duration, interval) = durations[duration_query]
|
2021-10-08 22:50:03 +00:00
|
|
|
if "queryStringParameters" in event and "uploader_callsign" in event["queryStringParameters"]:
|
2021-07-21 09:54:09 +00:00
|
|
|
interval = 1
|
|
|
|
lt = requested_time
|
|
|
|
gte = requested_time - timedelta(0, duration)
|
|
|
|
|
|
|
|
path = "listeners-*/_search"
|
|
|
|
payload = {
|
|
|
|
"timeout": "30s",
|
|
|
|
"aggs": {
|
|
|
|
"2": {
|
|
|
|
"terms": {
|
|
|
|
"field": "uploader_callsign.keyword",
|
|
|
|
"order": {"_key": "desc"},
|
|
|
|
"size": 10000,
|
|
|
|
},
|
|
|
|
"aggs": {
|
|
|
|
"3": {
|
|
|
|
"date_histogram": {
|
|
|
|
"field": "ts",
|
|
|
|
"fixed_interval": f"{str(interval)}s",
|
|
|
|
"min_doc_count": 1,
|
|
|
|
},
|
|
|
|
"aggs": {
|
|
|
|
"1": {
|
|
|
|
"top_hits": {
|
|
|
|
# "docvalue_fields": [
|
|
|
|
# {"field": "position"},
|
|
|
|
# {"field": "alt"},
|
|
|
|
# {"field": "datetime"},
|
|
|
|
# ],
|
|
|
|
# "_source": "position",
|
|
|
|
"size": 1,
|
|
|
|
"sort": [{"ts": {"order": "desc"}}],
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
}
|
|
|
|
},
|
|
|
|
}
|
|
|
|
},
|
|
|
|
"query": {
|
|
|
|
"bool": {
|
|
|
|
"filter": [
|
|
|
|
{"match_all": {}},
|
2021-10-08 22:50:03 +00:00
|
|
|
{"exists": { "field": "uploader_position"}},
|
2021-07-21 09:54:09 +00:00
|
|
|
{
|
|
|
|
"range": {
|
|
|
|
"ts": {"gte": gte.isoformat(), "lt": lt.isoformat()}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
]
|
|
|
|
}
|
|
|
|
},
|
|
|
|
}
|
|
|
|
if "queryStringParameters" in event:
|
|
|
|
if "uploader_callsign" in event["queryStringParameters"]:
|
|
|
|
payload["query"]["bool"]["filter"].append(
|
|
|
|
{
|
|
|
|
"match_phrase": {
|
|
|
|
"uploader_callsign": str(event["queryStringParameters"]["uploader_callsign"])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
)
|
|
|
|
results = es_request(payload, path, "POST")
|
|
|
|
output = {
|
|
|
|
sonde["key"]: {
|
|
|
|
data["key_as_string"]: data["1"]["hits"]["hits"][0]["_source"]
|
|
|
|
for data in sonde["3"]["buckets"]
|
|
|
|
}
|
|
|
|
for sonde in results["aggregations"]["2"]["buckets"]
|
|
|
|
}
|
|
|
|
|
|
|
|
compressed = BytesIO()
|
|
|
|
with gzip.GzipFile(fileobj=compressed, mode='w') as f:
|
|
|
|
json_response = json.dumps(output)
|
|
|
|
f.write(json_response.encode('utf-8'))
|
|
|
|
|
|
|
|
gzippedResponse = compressed.getvalue()
|
|
|
|
return {
|
|
|
|
"body": base64.b64encode(gzippedResponse).decode(),
|
|
|
|
"isBase64Encoded": True,
|
|
|
|
"statusCode": 200,
|
|
|
|
"headers": {
|
|
|
|
"Content-Encoding": "gzip",
|
|
|
|
"content-type": "application/json"
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
2021-03-28 05:00:44 +00:00
|
|
|
|
2021-10-04 10:55:51 +00:00
|
|
|
def get_sites(event, context):
|
|
|
|
|
|
|
|
path = "sites/_search"
|
|
|
|
payload = {
|
|
|
|
"version": True,
|
|
|
|
"size": 10000,
|
|
|
|
"_source": {
|
|
|
|
"excludes": []
|
|
|
|
},
|
|
|
|
"query": {
|
|
|
|
"bool": {
|
|
|
|
"filter": [
|
|
|
|
{
|
|
|
|
"match_all": {}
|
|
|
|
}
|
|
|
|
]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if "queryStringParameters" in event:
|
|
|
|
if "station" in event["queryStringParameters"]:
|
|
|
|
payload["query"]["bool"]["filter"].append(
|
|
|
|
{
|
|
|
|
"match_phrase": {
|
|
|
|
"station": str(event["queryStringParameters"]["station"])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
)
|
|
|
|
results = es_request(payload, path, "POST")
|
|
|
|
output = {x['_source']['station']: x['_source'] for x in results['hits']['hits']}
|
|
|
|
|
|
|
|
compressed = BytesIO()
|
|
|
|
with gzip.GzipFile(fileobj=compressed, mode='w') as f:
|
|
|
|
json_response = json.dumps(output)
|
|
|
|
f.write(json_response.encode('utf-8'))
|
|
|
|
|
|
|
|
gzippedResponse = compressed.getvalue()
|
|
|
|
return {
|
|
|
|
"body": base64.b64encode(gzippedResponse).decode(),
|
|
|
|
"isBase64Encoded": True,
|
|
|
|
"statusCode": 200,
|
|
|
|
"headers": {
|
|
|
|
"Content-Encoding": "gzip",
|
|
|
|
"content-type": "application/json"
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
2021-02-22 06:13:30 +00:00
|
|
|
|
2021-02-02 07:14:38 +00:00
|
|
|
|
|
|
|
def es_request(payload, path, method):
|
2021-02-02 10:44:39 +00:00
|
|
|
# get aws creds
|
2021-02-02 07:14:38 +00:00
|
|
|
session = boto3.Session()
|
|
|
|
|
|
|
|
params = json.dumps(payload)
|
2021-09-12 13:25:34 +00:00
|
|
|
compressed = BytesIO()
|
|
|
|
with gzip.GzipFile(fileobj=compressed, mode='w') as f:
|
|
|
|
f.write(params.encode('utf-8'))
|
|
|
|
params = compressed.getvalue()
|
|
|
|
|
|
|
|
headers = {"Host": HOST, "Content-Type": "application/json", "Content-Encoding":"gzip"}
|
2021-02-02 10:44:39 +00:00
|
|
|
request = AWSRequest(
|
|
|
|
method="POST", url=f"https://{HOST}/{path}", data=params, headers=headers
|
|
|
|
)
|
|
|
|
SigV4Auth(boto3.Session().get_credentials(), "es", "us-east-1").add_auth(request)
|
2021-12-11 05:17:41 +00:00
|
|
|
#p = Process(target=mirror, args=(path,params)).start()
|
2021-11-29 10:09:32 +00:00
|
|
|
r = http_session.send(request.prepare())
|
2021-02-02 07:14:38 +00:00
|
|
|
return json.loads(r.text)
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
2021-09-12 13:25:34 +00:00
|
|
|
#print(get_sondes({"queryStringParameters":{"lat":"-32.7933","lon":"151.8358","distance":"5000", "last":"604800"}}, {}))
|
2021-02-22 06:13:30 +00:00
|
|
|
# mode: 6hours
|
2021-03-28 05:00:44 +00:00
|
|
|
# type: positions
|
|
|
|
# format: json
|
|
|
|
# max_positions: 0
|
|
|
|
# position_id: 0
|
|
|
|
# vehicles: RS_*;*chase
|
2021-07-21 09:54:09 +00:00
|
|
|
# print(
|
|
|
|
# datanew(
|
|
|
|
# {
|
|
|
|
# "queryStringParameters": {
|
|
|
|
# "mode": "single",
|
|
|
|
# "format": "json",
|
|
|
|
# "position_id": "S1443103-2021-07-20T12:46:19.040000Z"
|
|
|
|
# }
|
|
|
|
# },
|
|
|
|
# {},
|
|
|
|
# )
|
|
|
|
# )
|
2021-11-29 10:09:32 +00:00
|
|
|
# print(get_sites({},{}))
|
|
|
|
print(
|
|
|
|
get_telem(
|
|
|
|
{
|
|
|
|
"queryStringParameters": {
|
2021-12-11 05:17:41 +00:00
|
|
|
"duration": "1d",
|
|
|
|
"serial": "S4430086"
|
2021-11-29 10:09:32 +00:00
|
|
|
}},{}
|
2021-10-04 08:08:54 +00:00
|
|
|
|
2021-11-29 10:09:32 +00:00
|
|
|
)
|
|
|
|
)
|
2021-07-24 09:33:29 +00:00
|
|
|
# print (
|
|
|
|
# get_chase(
|
|
|
|
# {"queryStringParameters": {
|
|
|
|
# "duration": "1d"
|
|
|
|
# }
|
|
|
|
# },
|
|
|
|
# {}
|
|
|
|
# )
|
|
|
|
# )
|
2021-07-20 13:46:44 +00:00
|
|
|
|
2021-07-21 09:54:09 +00:00
|
|
|
|
|
|
|
# print(
|
|
|
|
# datanew(
|
|
|
|
# {
|
|
|
|
# "queryStringParameters": {
|
|
|
|
# "type": "positions",
|
|
|
|
# "mode": "3hours",
|
2021-07-24 09:33:29 +00:00
|
|
|
# "position_id": "0"
|
2021-07-21 09:54:09 +00:00
|
|
|
# }
|
|
|
|
# },
|
|
|
|
# {},
|
|
|
|
# )
|
|
|
|
# )
|
2021-08-20 02:49:55 +00:00
|
|
|
# print(
|
|
|
|
# get_telem(
|
|
|
|
# {
|
|
|
|
# "queryStringParameters":{
|
|
|
|
# # "serial": "S3210639",
|
|
|
|
# "duration": "3h",
|
|
|
|
# # "datetime": "2021-07-26T06:49:29.001000Z"
|
|
|
|
# }
|
|
|
|
# }, {}
|
|
|
|
# )
|
|
|
|
# )
|
2021-07-21 09:54:09 +00:00
|
|
|
|