-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Datasources] Add: simple Google Analytics query runner #1405
Changes from 3 commits
de64c4c
0a0ca21
c94dace
8eefad2
92dee61
6bdc863
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,112 @@ | ||
from base64 import b64decode | ||
import json | ||
import logging | ||
from redash.query_runner import * | ||
from redash.utils import JSONEncoder | ||
from urlparse import urlparse, parse_qs | ||
from datetime import datetime | ||
logger = logging.getLogger(__name__) | ||
|
||
try: | ||
import gspread | ||
from oauth2client.client import SignedJwtAssertionCredentials | ||
from apiclient.discovery import build | ||
import httplib2 | ||
enabled = True | ||
except ImportError as e: | ||
logger.info(str(e)) | ||
enabled = False | ||
|
||
|
||
def _load_key(filename): | ||
with open(filename, "rb") as f: | ||
return json.loads(f.read()) | ||
|
||
|
||
types_conv = dict( | ||
STRING=TYPE_STRING, | ||
INTEGER=TYPE_INTEGER, | ||
FLOAT=TYPE_FLOAT, | ||
DATE=TYPE_DATE, | ||
DATETIME=TYPE_DATETIME | ||
) | ||
|
||
|
||
class GoogleAnalytics(BaseQueryRunner): | ||
@classmethod | ||
def annotate_query(cls): | ||
return False | ||
|
||
@classmethod | ||
def type(cls): | ||
return "google_analytics" | ||
|
||
@classmethod | ||
def enabled(cls): | ||
return enabled | ||
|
||
@classmethod | ||
def configuration_schema(cls): | ||
return { | ||
'type': 'object', | ||
'properties': { | ||
'jsonKeyFile': { | ||
"type": "string", | ||
'title': 'JSON Key File' | ||
} | ||
}, | ||
'required': ['jsonKeyFile'], | ||
'secret': ['jsonKeyFile'] | ||
} | ||
|
||
def __init__(self, configuration): | ||
super(GoogleAnalytics, self).__init__(configuration) | ||
|
||
def _get_analytics_service(self): | ||
scope = ['https://www.googleapis.com/auth/analytics.readonly'] | ||
key = json.loads(b64decode(self.configuration['jsonKeyFile'])) | ||
credentials = SignedJwtAssertionCredentials(key['client_email'], key["private_key"], scope=scope) | ||
return build('analytics', 'v3', http=credentials.authorize(httplib2.Http())) | ||
|
||
def run_query(self, query, user): | ||
logger.info("Analytics is about to execute query: %s", query) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
|
||
params = parse_qs(urlparse(query).query, keep_blank_values=True) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. How about we expect a JSON object instead of a query string? It will make parsing simpler, but also less error prone (most users won't know how to properly escape the query string). |
||
for key in params.keys(): | ||
params[key] = ','.join(params[key]) | ||
if '-' in key: | ||
params[key.replace('-', '_')] = params.pop(key) | ||
if len(params) > 0: | ||
response = self._get_analytics_service().data().ga().get(**params).execute() | ||
columns = [] | ||
for h in response['columnHeaders']: | ||
if h['name'] == 'ga:date': | ||
h['dataType'] = 'DATE' | ||
elif h['name'] == 'ga:dateHour': | ||
h['dataType'] = 'DATETIME' | ||
columns.append({ | ||
'name': h['name'], | ||
'friendly_name': h['name'].split(':', 1)[1], | ||
'type': types_conv.get(h['dataType'], 'string') | ||
}) | ||
rows = [] | ||
for r in response['rows']: | ||
d = {} | ||
for c, value in enumerate(r): | ||
column_name = response['columnHeaders'][c]['name'] | ||
column_type = filter(lambda col: col['name'] == column_name, columns)[0]['type'] | ||
if column_type == TYPE_DATE: | ||
value = datetime.strptime(value, '%Y%m%d') | ||
elif column_type == TYPE_DATETIME: | ||
if len(value) == 10: | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. and if not? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. There's only one TYPE_DATETIME column on it has 10 chars length. So I can add raise Exception if length is different. |
||
value = datetime.strptime(value, '%Y%m%d%H') | ||
d[column_name] = value | ||
rows.append(d) | ||
data = {'columns': columns, 'rows': rows} | ||
error = None | ||
json_data = json.dumps(data, cls=JSONEncoder) | ||
else: | ||
error = 'Wrong query format' | ||
json_data = None | ||
return json_data, error | ||
|
||
register(GoogleAnalytics) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
you don't need gspread here.