X-Git-Url: http://git.maemo.org/git/?p=theonering;a=blobdiff_plain;f=src%2Fgvoice%2Fconversations.py;h=7cad027581ee438c65edaa1ba396ecea322c4dfc;hp=5f3d371219fc44646693a110e7c9e6f3a42f1481;hb=76d24c8d4a16fe0e620e8d26b676876a19c50162;hpb=dc433fea2a937e651729013ac30341b83176f168 diff --git a/src/gvoice/conversations.py b/src/gvoice/conversations.py index 5f3d371..7cad027 100644 --- a/src/gvoice/conversations.py +++ b/src/gvoice/conversations.py @@ -14,19 +14,27 @@ except ImportError: import constants import util.coroutines as coroutines import util.misc as misc_utils +import util.go_utils as gobject_utils _moduleLogger = logging.getLogger(__name__) +class ConversationError(RuntimeError): + + pass + + class Conversations(object): OLDEST_COMPATIBLE_FORMAT_VERSION = misc_utils.parse_version("0.8.0") - OLDEST_MESSAGE_WINDOW = datetime.timedelta(days=60) - def __init__(self, getter): + def __init__(self, getter, asyncPool): self._get_raw_conversations = getter + self._asyncPool = asyncPool self._conversations = {} + self._loadedFromCache = False + self._hasDoneUpdate = False self.updateSignalHandler = coroutines.CoTee() @@ -35,6 +43,7 @@ class Conversations(object): return repr(self._get_raw_conversations.__name__) def load(self, path): + _moduleLogger.debug("%s Loading cache" % (self._name, )) assert not self._conversations try: with open(path, "rb") as f: @@ -43,11 +52,13 @@ class Conversations(object): _moduleLogger.exception("While loading for %s" % self._name) return - if misc_utils.compare_versions( + if convs and misc_utils.compare_versions( self.OLDEST_COMPATIBLE_FORMAT_VERSION, misc_utils.parse_version(fileVersion), ) <= 0: + _moduleLogger.info("%s Loaded cache" % (self._name, )) self._conversations = convs + self._loadedFromCache = True else: _moduleLogger.debug( "%s Skipping cache due to version mismatch (%s-%s)" % ( @@ -56,23 +67,42 @@ class Conversations(object): ) def save(self, path): + _moduleLogger.info("%s Saving cache" % (self._name, )) + if not self._conversations: + _moduleLogger.info("%s Odd, no conversations to cache. Did we never load the cache?" % (self._name, )) + return + try: - for conv in self._conversations.itervalues(): - conv.compress(self.OLDEST_MESSAGE_WINDOW) dataToDump = (constants.__version__, constants.__build__, self._conversations) with open(path, "wb") as f: pickle.dump(dataToDump, f, pickle.HIGHEST_PROTOCOL) except (pickle.PickleError, IOError): _moduleLogger.exception("While saving for %s" % self._name) + _moduleLogger.info("%s Cache saved" % (self._name, )) def update(self, force=False): if not force and self._conversations: return + le = gobject_utils.AsyncLinearExecution(self._asyncPool, self._update) + le.start() + + @misc_utils.log_exception(_moduleLogger) + def _update(self): + try: + conversationResult = yield ( + self._get_raw_conversations, + (), + {}, + ) + except Exception: + _moduleLogger.exception("%s While updating conversations" % (self._name, )) + return + oldConversationIds = set(self._conversations.iterkeys()) updateConversationIds = set() - conversations = list(self._get_raw_conversations()) + conversations = list(conversationResult) conversations.sort() for conversation in conversations: key = misc_utils.normalize_number(conversation.number) @@ -82,20 +112,39 @@ class Conversations(object): mergedConversations = MergedConversations() self._conversations[key] = mergedConversations + if self._loadedFromCache or self._hasDoneUpdate: + markAllAsRead = False + else: + markAllAsRead = True + try: - mergedConversations.append_conversation(conversation) + mergedConversations.append_conversation(conversation, markAllAsRead) isConversationUpdated = True + except ConversationError, e: + isConversationUpdated = False + except AssertionError, e: + _moduleLogger.debug("%s Skipping conversation for %r because '%s'" % (self._name, key, e)) + isConversationUpdated = False except RuntimeError, e: - if False: - _moduleLogger.debug("%s Skipping conversation for %r because '%s'" % (self._name, key, e)) + _moduleLogger.debug("%s Skipping conversation for %r because '%s'" % (self._name, key, e)) isConversationUpdated = False if isConversationUpdated: updateConversationIds.add(key) + for key in updateConversationIds: + mergedConv = self._conversations[key] + _moduleLogger.debug("%s \tUpdated %s" % (self._name, key)) + for conv in mergedConv.conversations: + message = "%s \t\tUpdated %s (%r) %r %r %r" % ( + self._name, conv.id, conv.time, conv.isRead, conv.isArchived, len(conv.messages) + ) + _moduleLogger.debug(message) + if updateConversationIds: message = (self, updateConversationIds, ) self.updateSignalHandler.stage.send(message) + self._hasDoneUpdate = True def get_conversations(self): return self._conversations.iterkeys() @@ -118,24 +167,32 @@ class MergedConversations(object): def __init__(self): self._conversations = [] - def append_conversation(self, newConversation): + def append_conversation(self, newConversation, markAllAsRead): self._validate(newConversation) - similarExist = False for similarConversation in self._find_related_conversation(newConversation.id): self._update_previous_related_conversation(similarConversation, newConversation) self._remove_repeats(similarConversation, newConversation) - similarExist = True - if similarExist: - # Hack to reduce a race window with GV marking messages as read - # because it thinks we replied when really we replied to the - # previous message. Clients of this code are expected to handle - # this gracefully. Other race conditions may exist but clients are - # responsible for them - if newConversation.messages: - newConversation.isRead = False - else: + + # HACK: Because GV marks all messages as read when you reply it has + # the following race: + # 1. Get all messages + # 2. Contact sends a text + # 3. User sends a text marking contacts text as read + # 4. Get all messages not returning text from step 2 + # This isn't a problem for voicemails but we don't know(?( enough. + # So we hack around this by: + # * We cache to disk the history of messages sent/received + # * On first run we mark all server messages as read due to no cache + # * If not first load or from cache (disk or in-memory) then it must be unread + if newConversation.type != newConversation.TYPE_VOICEMAIL: + if markAllAsRead: newConversation.isRead = True - self._conversations.append(newConversation) + else: + newConversation.isRead = False + + if newConversation.messages: + # must not have had all items removed due to duplicates + self._conversations.append(newConversation) def to_dict(self): selfDict = {} @@ -146,21 +203,6 @@ class MergedConversations(object): def conversations(self): return self._conversations - def compress(self, timedelta): - now = datetime.datetime.now() - oldNumConvs = len(self._conversations) - oldConvs = self._conversations - self._conversations = [ - conv - for conv in self._conversations - if (now - conv.time) < timedelta - ] - newNumConvs = len(self._conversations) - if oldNumConvs != newNumConvs: - _moduleLogger.debug("Compressed conversations from %s to %s" % (oldNumConvs, newNumConvs)) - else: - _moduleLogger.debug("Did not compress, %s" % (newNumConvs)) - def _validate(self, newConversation): if not self._conversations: return @@ -172,7 +214,7 @@ class MergedConversations(object): ) if newConversation.time <= self._conversations[-1].time: - raise RuntimeError("Conversations got out of order") + raise ConversationError("Conversations got out of order") def _find_related_conversation(self, convId): similarConversations = ( @@ -245,3 +287,18 @@ class FilterOutReported(object): if filteredConversations and self._lastMessageTimestamp < filteredConversations[0].time: self._lastMessageTimestamp = filteredConversations[0].time return filteredConversations + + +def print_conversations(path): + import pprint + + try: + with open(path, "rb") as f: + fileVersion, fileBuild, convs = pickle.load(f) + except (pickle.PickleError, IOError, EOFError, ValueError): + _moduleLogger.exception("") + else: + for key, value in convs.iteritems(): + convs[key] = value.to_dict() + pprint.pprint((fileVersion, fileBuild)) + pprint.pprint(convs)