2020-02-19 22:12:56 +00:00
|
|
|
import atexit
|
2021-11-17 02:12:22 +00:00
|
|
|
import base64
|
|
|
|
import json
|
|
|
|
import os
|
2020-02-19 22:12:56 +00:00
|
|
|
from collections.abc import MutableMapping
|
2021-11-17 02:12:22 +00:00
|
|
|
from datetime import datetime, date, timedelta
|
|
|
|
from threading import Thread, Event
|
|
|
|
from typing import List
|
2020-02-19 22:12:56 +00:00
|
|
|
|
|
|
|
from PySide2 import QtCore
|
|
|
|
|
|
|
|
data_dir_path = os.path.join(QtCore.QStandardPaths.writableLocation(QtCore.QStandardPaths.AppDataLocation),
|
|
|
|
"fimefracking")
|
|
|
|
|
|
|
|
tasks_path = os.path.join(data_dir_path, "tasks.json")
|
|
|
|
|
|
|
|
data_path = os.path.join(data_dir_path, "data_{}.json")
|
2020-02-24 17:19:01 +00:00
|
|
|
save_delay = 3 * 60
|
2021-11-17 02:12:22 +00:00
|
|
|
max_jira_tasks = 50
|
2020-02-19 22:12:56 +00:00
|
|
|
|
|
|
|
|
|
|
|
class Tasks:
|
2021-11-17 02:12:22 +00:00
|
|
|
def __init__(self, data):
|
|
|
|
self._data = data
|
|
|
|
if "tasks" in self._data:
|
|
|
|
self._tasks = list(map(lambda x: base64.b64decode(x.encode("utf-8")).decode("utf-8"), self._data["tasks"]))
|
2020-02-19 22:12:56 +00:00
|
|
|
else:
|
|
|
|
self._tasks = []
|
2021-11-17 02:12:22 +00:00
|
|
|
if "jira_tasks" in self._data:
|
|
|
|
self._jira_tasks_usage = dict()
|
|
|
|
for k, v in self._data["jira_tasks"].items():
|
|
|
|
key = base64.b64decode(k.encode("utf-8")).decode("utf-8")
|
|
|
|
self._jira_tasks_usage[key] = datetime.fromisoformat(v)
|
|
|
|
self._jira_tasks = sorted(self._jira_tasks_usage.keys(), key=lambda x: self._jira_tasks_usage[x])
|
|
|
|
else:
|
|
|
|
self._jira_tasks_usage = dict()
|
|
|
|
self._jira_tasks = []
|
2020-02-19 22:12:56 +00:00
|
|
|
|
|
|
|
@property
|
2021-11-17 02:12:22 +00:00
|
|
|
def tasks(self) -> List[str]:
|
2020-02-19 22:12:56 +00:00
|
|
|
return self._tasks
|
|
|
|
|
|
|
|
@tasks.setter
|
|
|
|
def tasks(self, tasks):
|
|
|
|
self._tasks = tasks
|
2020-02-20 14:22:40 +00:00
|
|
|
encoded_tasks = list(map(lambda x: base64.b64encode(x.encode("utf-8")).decode("utf-8"), self._tasks))
|
2021-11-17 02:12:22 +00:00
|
|
|
self._data["tasks"] = encoded_tasks
|
|
|
|
|
|
|
|
@property
|
|
|
|
def jira_tasks(self):
|
|
|
|
return self._jira_tasks
|
|
|
|
|
|
|
|
def add_jira_task(self, task_name):
|
|
|
|
self._jira_tasks.append(task_name)
|
|
|
|
self._jira_tasks_usage[task_name] = datetime.now()
|
|
|
|
if len(self._jira_tasks_usage) > max_jira_tasks:
|
|
|
|
sorted_tasks = sorted(self._jira_tasks_usage.keys(), key=lambda x: self._jira_tasks_usage[x])
|
|
|
|
overhang_tasks = sorted_tasks[:len(sorted_tasks) - max_jira_tasks]
|
|
|
|
for task in overhang_tasks:
|
|
|
|
del self._jira_tasks_usage[task]
|
|
|
|
self._save_jira_tasks()
|
|
|
|
|
|
|
|
def update_jira_task_usage(self, task_name):
|
|
|
|
if task_name in self._jira_tasks_usage:
|
|
|
|
self._jira_tasks_usage[task_name] = datetime.now()
|
|
|
|
self._save_jira_tasks()
|
|
|
|
|
|
|
|
def _save_jira_tasks(self):
|
|
|
|
serialized = dict()
|
|
|
|
for k, v in self._jira_tasks_usage.items():
|
|
|
|
key = base64.b64encode(k.encode("utf-8")).decode("utf-8")
|
|
|
|
serialized[key] = datetime.isoformat(v)
|
|
|
|
self._data["jira_tasks"] = serialized
|
|
|
|
|
|
|
|
@property
|
|
|
|
def all_tasks(self):
|
|
|
|
return self.tasks + self.jira_tasks
|
2020-02-19 22:12:56 +00:00
|
|
|
|
|
|
|
|
|
|
|
class Data(MutableMapping):
|
|
|
|
def __init__(self):
|
|
|
|
if not os.path.exists(data_dir_path):
|
|
|
|
os.mkdir(data_dir_path)
|
|
|
|
self._cache = {}
|
2020-02-25 19:20:37 +00:00
|
|
|
self._hot_keys = set()
|
2020-02-20 14:22:40 +00:00
|
|
|
self._trunning = False
|
2020-02-19 22:12:56 +00:00
|
|
|
self._tevent = Event()
|
|
|
|
self._thread = None
|
|
|
|
|
|
|
|
def cleanup():
|
2020-02-20 14:22:40 +00:00
|
|
|
self._trunning = False
|
2020-02-19 22:12:56 +00:00
|
|
|
self._tevent.set()
|
|
|
|
if self._thread:
|
|
|
|
self._thread.join()
|
2020-02-20 14:22:40 +00:00
|
|
|
|
2020-02-19 22:12:56 +00:00
|
|
|
atexit.register(cleanup)
|
|
|
|
|
|
|
|
def __getitem__(self, key):
|
|
|
|
dpath = data_path.format(key)
|
|
|
|
if key not in self._cache and os.path.exists(dpath):
|
|
|
|
with open(dpath, "r") as f:
|
2020-02-20 14:22:40 +00:00
|
|
|
self._cache[key] = json.loads(f.read())
|
2020-02-19 22:12:56 +00:00
|
|
|
return self._cache[key]
|
|
|
|
|
|
|
|
def __setitem__(self, key, value):
|
|
|
|
self._cache[key] = value
|
2020-02-25 19:20:37 +00:00
|
|
|
self._hot_keys.add(key)
|
2020-02-19 22:12:56 +00:00
|
|
|
self._schedule_save()
|
|
|
|
|
|
|
|
def _schedule_save(self):
|
2020-02-20 14:22:40 +00:00
|
|
|
if self._trunning:
|
2020-02-19 22:12:56 +00:00
|
|
|
return
|
2020-02-20 14:22:40 +00:00
|
|
|
self._trunning = True
|
2020-02-19 22:12:56 +00:00
|
|
|
self._thread = Thread(target=self._executor, daemon=True)
|
|
|
|
self._thread.start()
|
|
|
|
|
|
|
|
def _executor(self):
|
2020-02-20 14:22:40 +00:00
|
|
|
while self._trunning:
|
2020-02-19 22:12:56 +00:00
|
|
|
self._tevent.wait(save_delay)
|
|
|
|
self._save()
|
|
|
|
|
|
|
|
def _save(self):
|
|
|
|
for key in self._hot_keys:
|
|
|
|
print(f"... saving dict {key} ...")
|
|
|
|
to_write = self._cache[key] # apparently thread-safe
|
|
|
|
with open(data_path.format(key), "w+") as f:
|
|
|
|
f.write(json.dumps(to_write))
|
2020-02-25 19:20:37 +00:00
|
|
|
self._hot_keys = set()
|
2020-02-19 22:12:56 +00:00
|
|
|
self._saving = False
|
|
|
|
|
|
|
|
def __delitem__(self, key):
|
2020-02-20 14:22:40 +00:00
|
|
|
return NotImplemented
|
2020-02-19 22:12:56 +00:00
|
|
|
|
|
|
|
def __iter__(self):
|
2020-02-20 14:22:40 +00:00
|
|
|
return NotImplemented
|
2020-02-19 22:12:56 +00:00
|
|
|
|
|
|
|
def __len__(self):
|
2020-02-20 14:22:40 +00:00
|
|
|
# TODO use glob?
|
|
|
|
return NotImplemented
|
2020-02-19 22:12:56 +00:00
|
|
|
|
|
|
|
def __repr__(self):
|
|
|
|
return f"{type(self).__name__}({self._cache})"
|
2020-02-20 14:22:40 +00:00
|
|
|
|
|
|
|
|
|
|
|
class Log:
|
2021-11-17 02:12:22 +00:00
|
|
|
def __init__(self, data):
|
|
|
|
self._data = data
|
2020-02-20 14:22:40 +00:00
|
|
|
|
2020-02-24 15:54:28 +00:00
|
|
|
def cleanup():
|
|
|
|
self.log("End")
|
|
|
|
|
|
|
|
atexit.register(cleanup)
|
|
|
|
|
2020-02-24 12:16:17 +00:00
|
|
|
def log(self, task, ptime=None):
|
|
|
|
if ptime is None:
|
|
|
|
ptime = datetime.now()
|
2020-02-24 14:41:08 +00:00
|
|
|
# round to nearest minute
|
|
|
|
round_min = timedelta(minutes=round(ptime.second/60))
|
|
|
|
ptime = ptime - timedelta(seconds=ptime.second) + round_min
|
|
|
|
# month dance necessary to trigger Data.__setitem__
|
2020-02-20 14:22:40 +00:00
|
|
|
month = self._data.setdefault(ptime.strftime("%Y-%m"), {})
|
|
|
|
month.setdefault(ptime.strftime("%d"), [])\
|
2020-02-24 14:41:08 +00:00
|
|
|
.append(f"{ptime.strftime('%H:%M')} {base64.b64encode(task.encode('utf-8')).decode('utf-8')}")
|
|
|
|
self._data[ptime.strftime("%Y-%m")] = month
|
2020-02-20 14:22:40 +00:00
|
|
|
|
2020-02-24 15:54:28 +00:00
|
|
|
def last_log(self, pdate=None):
|
|
|
|
if pdate is None:
|
|
|
|
pdate = date.today()
|
|
|
|
if pdate.strftime("%Y-%m") not in self._data \
|
|
|
|
or pdate.strftime("%d") not in self._data[pdate.strftime("%Y-%m")] \
|
|
|
|
or len(self._data[pdate.strftime("%Y-%m")][pdate.strftime("%d")]) == 0:
|
2020-02-20 14:22:40 +00:00
|
|
|
return None
|
2020-02-24 15:54:28 +00:00
|
|
|
last = base64.b64decode(
|
2020-02-20 14:22:40 +00:00
|
|
|
self._data[pdate.strftime("%Y-%m")][pdate.strftime("%d")][-1].split()[1].encode("utf-8")).decode("utf-8")
|
2020-02-24 15:54:28 +00:00
|
|
|
if last == "End":
|
|
|
|
month = self._data[pdate.strftime("%Y-%m")]
|
|
|
|
del month[pdate.strftime("%d")][-1]
|
|
|
|
self._data[pdate.strftime("%Y-%m")] = month
|
2020-02-25 19:20:37 +00:00
|
|
|
if len(self._data[pdate.strftime("%Y-%m")][pdate.strftime("%d")]) == 0:
|
|
|
|
return None
|
2020-02-24 15:54:28 +00:00
|
|
|
last = base64.b64decode(
|
|
|
|
self._data[pdate.strftime("%Y-%m")][pdate.strftime("%d")][-1].split()[1].encode("utf-8")).decode("utf-8")
|
|
|
|
return last
|
2020-02-20 14:22:40 +00:00
|
|
|
|
2020-02-24 14:41:08 +00:00
|
|
|
def report(self, pdate=None):
|
|
|
|
if pdate is None:
|
|
|
|
pdate = date.today()
|
|
|
|
return Report(self._data, pdate)
|
|
|
|
|
|
|
|
|
|
|
|
class Report:
|
|
|
|
def __init__(self, data, pdate):
|
|
|
|
self._data = data
|
|
|
|
self._date = pdate
|
2020-03-03 22:20:36 +00:00
|
|
|
self._sum_len = 0
|
2020-03-02 20:53:19 +00:00
|
|
|
self._prev = None
|
|
|
|
self._next = None
|
|
|
|
self._update_prev_next()
|
2020-02-24 14:41:08 +00:00
|
|
|
|
|
|
|
def report(self):
|
2020-02-20 14:22:40 +00:00
|
|
|
tmp = []
|
2020-02-25 19:20:37 +00:00
|
|
|
if self._date.strftime("%Y-%m") in self._data \
|
|
|
|
and self._date.strftime("%d") in self._data[self._date.strftime("%Y-%m")]:
|
|
|
|
for e in self._data[self._date.strftime("%Y-%m")][self._date.strftime("%d")]:
|
|
|
|
tstr, b64str = e.split()
|
|
|
|
task = base64.b64decode(b64str.encode("utf-8")).decode("utf-8")
|
|
|
|
start_time = datetime.combine(self._date, datetime.strptime(tstr, "%H:%M").time())
|
|
|
|
tmp.append((task, start_time))
|
2020-02-24 15:54:28 +00:00
|
|
|
if self._date == date.today():
|
|
|
|
tmp.append(("End", datetime.now()))
|
2020-02-20 14:22:40 +00:00
|
|
|
|
|
|
|
ret = []
|
2020-03-02 18:43:55 +00:00
|
|
|
tasks_sums = {}
|
|
|
|
total_sum = timedelta()
|
2020-02-20 14:22:40 +00:00
|
|
|
for i, t in enumerate(tmp):
|
|
|
|
task, start_time = t
|
|
|
|
if i < len(tmp) - 1:
|
|
|
|
end_time = tmp[i+1][1]
|
2020-02-24 15:54:28 +00:00
|
|
|
duration = end_time - start_time
|
2020-02-26 22:13:47 +00:00
|
|
|
if task != "Pause":
|
2020-03-02 18:43:55 +00:00
|
|
|
task_sum = tasks_sums.setdefault(task, timedelta())
|
|
|
|
task_sum += duration
|
|
|
|
tasks_sums[task] = task_sum
|
|
|
|
total_sum += duration
|
2020-02-24 15:54:28 +00:00
|
|
|
dhours, rem = divmod(duration.seconds, 3600)
|
|
|
|
dmins, _ = divmod(rem, 60)
|
2020-02-24 16:15:34 +00:00
|
|
|
ret.append([task, start_time.strftime("%H:%M"), f"{dhours:02d}:{dmins:02d}"])
|
2020-02-20 14:22:40 +00:00
|
|
|
else:
|
2020-02-24 16:15:34 +00:00
|
|
|
ret.append([task, start_time.strftime("%H:%M"), ""])
|
2020-02-24 12:16:17 +00:00
|
|
|
|
2020-02-24 16:15:34 +00:00
|
|
|
ret.append(["", "", ""])
|
2020-03-02 18:43:55 +00:00
|
|
|
ret.append(["", "Sums", ""])
|
|
|
|
for k, v in tasks_sums.items():
|
|
|
|
dhours, rem = divmod(v.seconds, 3600)
|
|
|
|
dmins, _ = divmod(rem, 60)
|
|
|
|
ret.append([k, "", f"{dhours:02d}:{dmins:02d}"])
|
|
|
|
dhours, rem = divmod(total_sum.seconds, 3600)
|
2020-02-24 12:16:17 +00:00
|
|
|
dmins, _ = divmod(rem, 60)
|
2020-03-02 18:43:55 +00:00
|
|
|
ret.append(["Total sum", "", f"{dhours:02d}:{dmins:02d}"])
|
2020-03-03 22:20:36 +00:00
|
|
|
self._sum_len = 3 + len(tasks_sums)
|
|
|
|
if self._date == date.today():
|
|
|
|
self._sum_len += 1
|
|
|
|
return ret, len(ret) - (4 + len(tasks_sums))
|
2020-02-24 14:41:08 +00:00
|
|
|
|
|
|
|
def save(self, report):
|
2020-03-03 22:20:36 +00:00
|
|
|
report = report[:-self._sum_len]
|
2020-02-25 19:20:37 +00:00
|
|
|
if not report:
|
|
|
|
return
|
2020-02-24 14:41:08 +00:00
|
|
|
save_list = []
|
|
|
|
for tstr, ttime, _ in report:
|
|
|
|
b64str = base64.b64encode(tstr.encode("utf-8")).decode("utf-8")
|
|
|
|
save_string = f"{ttime} {b64str}"
|
|
|
|
save_list.append(save_string)
|
|
|
|
# month dance necessary to trigger Data.__setitem__
|
|
|
|
month = self._data[self._date.strftime("%Y-%m")]
|
|
|
|
if month[self._date.strftime("%d")] == save_list: # no changes
|
|
|
|
return
|
|
|
|
month[self._date.strftime("%d")] = save_list
|
|
|
|
self._data[self._date.strftime("%Y-%m")] = month
|
|
|
|
|
2020-03-02 20:53:19 +00:00
|
|
|
def _update_prev_next(self):
|
|
|
|
self._prev = None
|
|
|
|
self._next = None
|
|
|
|
for i in range(1, 32):
|
|
|
|
new_date = self._date - timedelta(days=i)
|
|
|
|
if new_date.strftime("%Y-%m") not in self._data:
|
|
|
|
break
|
|
|
|
if new_date.strftime("%d") in self._data[new_date.strftime("%Y-%m")]:
|
|
|
|
self._prev = new_date
|
|
|
|
break
|
|
|
|
for i in range(1, 32):
|
|
|
|
new_date = self._date + timedelta(days=i)
|
|
|
|
if new_date > date.today():
|
|
|
|
break
|
|
|
|
if new_date.strftime("%Y-%m") not in self._data:
|
|
|
|
break
|
|
|
|
if new_date.strftime("%d") in self._data[new_date.strftime("%Y-%m")]:
|
|
|
|
self._next = new_date
|
|
|
|
break
|
|
|
|
|
2020-02-24 14:41:08 +00:00
|
|
|
def prev_next_avail(self):
|
2020-03-02 20:53:19 +00:00
|
|
|
return self._prev is not None, self._next is not None
|
2020-02-24 14:41:08 +00:00
|
|
|
|
|
|
|
def previous(self):
|
2020-03-02 20:53:19 +00:00
|
|
|
self._date = self._prev
|
|
|
|
self._update_prev_next()
|
2020-02-24 14:41:08 +00:00
|
|
|
|
|
|
|
def next(self):
|
2020-03-02 20:53:19 +00:00
|
|
|
self._date = self._next
|
|
|
|
self._update_prev_next()
|
2020-02-24 17:18:35 +00:00
|
|
|
|
|
|
|
def date(self):
|
|
|
|
return self._date.strftime("%Y-%m-%d")
|