From a5a4cea9794406b51d79c4c841b05a58709e4559 Mon Sep 17 00:00:00 2001 From: Mattias Andrée Date: Sun, 6 Apr 2014 05:37:47 +0200 Subject: how did I not catch this regression bug before now MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: Mattias Andrée --- src/__main__.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) (limited to 'src') diff --git a/src/__main__.py b/src/__main__.py index e5b3203..cf4d619 100755 --- a/src/__main__.py +++ b/src/__main__.py @@ -433,7 +433,7 @@ def continuous_run(): with_fadein = lambda : (fadein_steps > 0) and (fadein_time is not None) and not panicgate with_fadeout = lambda : (fadeout_steps > 0) and (fadeout_time is not None) - try: + try: # TODO why does it look like there is a small jump at the beginning? ## Run until we get a signal to exit # When the program start we are fading in, # than we run in normal periodical state. @@ -456,7 +456,7 @@ def continuous_run(): # If we are using fading, if with_fadein(): # and just started - if trans_alpha == 0: + if trans_alpha == 1: # Apply fully clean adjustments, p(now(), 1 - trans_alpha) # and and sleep for a short period. @@ -466,7 +466,7 @@ def continuous_run(): # If we are not fading, which might actually # have beend should from `periodically`, which # is invoked by `p`, - if not with_fadein(): + if (not with_fadein()) or (trans_alpha < 0): # The jump to adjusted state and # stop transitioning trans_alpha = trans_delta = 0 -- cgit v1.2.3-70-g09d2