Logging notification tweaks

This commit is contained in:
kay.one 2011-07-05 23:17:21 -07:00
parent 46f90255bf
commit fc0d7fc72f
8 changed files with 29 additions and 28 deletions

View File

@ -136,7 +136,7 @@ namespace NzbDrone.Core
return parsedEpisode;
}
}
Logger.Warn("Unable to parse text into episode info. {0}", title);
Logger.Warn("Unable to parse episode info. {0}", title);
return null;
}

View File

@ -135,7 +135,7 @@ namespace NzbDrone.Core.Providers
{
ep.EpisodeFileId = fileId;
_episodeProvider.UpdateEpisode(ep);
Logger.Debug("Linking file {0} to {1}", filePath, ep);
Logger.Debug("Linking [{0}] > [{1}]", filePath, ep);
}
@ -156,7 +156,7 @@ namespace NzbDrone.Core.Providers
_diskProvider.CreateDirectory(newFile.DirectoryName);
//Do the rename
Logger.Debug("Moving file [{0}] > [{1}]", episodeFile.Path, newFile.FullName);
Logger.Debug("Moving [{0}] > [{1}]", episodeFile.Path, newFile.FullName);
_diskProvider.MoveFile(episodeFile.Path, newFile.FullName);
//Update the filename in the DB
@ -176,7 +176,7 @@ namespace NzbDrone.Core.Providers
{
if (!_diskProvider.FileExists(episodeFile.Path))
{
Logger.Trace("File {0} no longer exists on disk. removing from database.", episodeFile.Path);
Logger.Trace("File [{0}] no longer exists on disk. removing from db", episodeFile.Path);
//Set the EpisodeFileId for each episode attached to this file to 0
foreach (var episode in _episodeProvider.GetEpisodesByFileId(episodeFile.EpisodeFileId))

View File

@ -44,7 +44,7 @@ namespace NzbDrone.Core.Providers.Jobs
_seriesProvider.DeleteSeries(seriesId);
notification.CurrentMessage = String.Format("Successfully deleted '{0}'", title);
notification.CurrentMessage = String.Format("Successfully deleted '{0}' from database", title);
}
catch (Exception e)
{

View File

@ -53,11 +53,11 @@ namespace NzbDrone.Core.Providers.Jobs
{
notification.CurrentMessage = string.Format("Scanning disk for '{0}'", series.Title);
_diskScanProvider.Scan(series);
notification.CurrentMessage = string.Format("Media File Scan completed for '{0}'", series.Title);
notification.CurrentMessage = string.Format("Disk Scan completed for '{0}'", series.Title);
}
catch (Exception e)
{
Logger.ErrorException("An error has occured while scanning " + series.Title, e);
Logger.ErrorException("An error has occurred while scanning " + series.Title, e);
}
}
}

View File

@ -73,6 +73,8 @@ namespace NzbDrone.Core.Providers.Jobs
var updatedSeries = _seriesProvider.GetSeries(currentSeries.SeriesId);
AutoIgnoreSeasons(updatedSeries.SeriesId);
notification.CurrentMessage = String.Format("'{0}' was successfully imported", currentSeries.Title);
}
catch (Exception e)
{

View File

@ -73,7 +73,7 @@ namespace NzbDrone.Core.Providers.Jobs
{
if (_isRunning)
{
Logger.Info("Another instance of this job is already running. Ignoring request.");
Logger.Trace("Queue is already running. Ignoring scheduler's request.");
return false;
}
_isRunning = true;
@ -112,19 +112,19 @@ namespace NzbDrone.Core.Providers.Jobs
/// <remarks>Job is only added to the queue if same job with the same targetId doesn't already exist in the queue.</remarks>
public virtual bool QueueJob(Type jobType, int targetId = 0)
{
Logger.Debug("Adding job ({0}:{1}) to the queue", jobType, targetId);
Logger.Debug("Adding [{0}:{1}] to the queue", jobType.Name, targetId);
lock (Queue)
{
var queueTuple = new Tuple<Type, int>(jobType, targetId);
if (Queue.Contains(queueTuple))
{
Logger.Info("Job ({0}:{1}) already exists in queue. Skipping.", jobType, targetId);
Logger.Info("[{0}:{1}] already exists in job queue. Skipping.", jobType.Name, targetId);
return false;
}
Queue.Add(queueTuple);
Logger.Trace("Job ({0}:{1}) added to the queue", jobType, targetId);
Logger.Trace("Job [{0}:{1}] added to the queue", jobType.Name, targetId);
}
@ -132,7 +132,7 @@ namespace NzbDrone.Core.Providers.Jobs
{
if (_isRunning)
{
Logger.Trace("Queue is already running. Ignoring request.");
Logger.Trace("Queue is already running. No need to start it up.");
return true;
}
_isRunning = true;
@ -150,7 +150,7 @@ namespace NzbDrone.Core.Providers.Jobs
}
catch (Exception e)
{
Logger.ErrorException("Error has occured in queue processor thread", e);
Logger.ErrorException("Error has occurred in queue processor thread", e);
}
finally
{
@ -158,13 +158,13 @@ namespace NzbDrone.Core.Providers.Jobs
}
};
_jobThread = new Thread(starter) { Name = "JobQueueThread", Priority = ThreadPriority.BelowNormal };
_jobThread = new Thread(starter) { Name = "JobQueueThread" };
_jobThread.Start();
}
else
{
Logger.Warn("Execution lock has has fucked up. Thread still active. Ignoring request.");
Logger.Error("Execution lock has fucked up. Thread still active. Ignoring request.");
return true;
}
@ -230,7 +230,7 @@ namespace NzbDrone.Core.Providers.Jobs
var jobImplementation = _jobs.Where(t => t.GetType() == jobType).FirstOrDefault();
if (jobImplementation == null)
{
Logger.Error("Unable to locate implementation for '{0}'. Make sure its properly registered.", jobType.ToString());
Logger.Error("Unable to locate implementation for '{0}'. Make sure it is properly registered.", jobType);
return;
}
@ -240,7 +240,7 @@ namespace NzbDrone.Core.Providers.Jobs
{
try
{
Logger.Debug("Starting job '{0}'. Last execution {1}", settings.Name, settings.LastExecution);
Logger.Debug("Starting '{0}' job. Last execution {1}", settings.Name, settings.LastExecution);
var sw = Stopwatch.StartNew();
@ -252,12 +252,12 @@ namespace NzbDrone.Core.Providers.Jobs
settings.Success = true;
sw.Stop();
Logger.Debug("Job '{0}' successfully completed in {1}.{2} seconds.", jobImplementation.Name, sw.Elapsed.Seconds,sw.Elapsed.Milliseconds/100,
Logger.Debug("Job '{0}' successfully completed in {1}.{2} seconds.", jobImplementation.Name, sw.Elapsed.Seconds, sw.Elapsed.Milliseconds / 100,
sw.Elapsed.Seconds);
}
catch (Exception e)
{
Logger.ErrorException("An error has occurred while executing timer job " + jobImplementation.Name, e);
Logger.ErrorException("An error has occurred while executing job " + jobImplementation.Name, e);
_notification.Status = ProgressNotificationStatus.Failed;
_notification.CurrentMessage = jobImplementation.Name + " Failed.";

View File

@ -41,18 +41,17 @@ namespace NzbDrone.Core.Providers.Jobs
public virtual void Start(ProgressNotification notification, int targetId)
{
Logger.Debug("Starting New Download Scan Job");
var dropFolder = _configProvider.SabDropDirectory;
if (String.IsNullOrWhiteSpace(dropFolder))
{
Logger.Debug("Skipping drop folder scan. No drop folder is defined.");
Logger.Debug("No drop folder is defined. Skipping.");
return;
}
if (!_diskProvider.FolderExists(dropFolder))
{
Logger.Warn("Unable to Scan for New Downloads - folder Doesn't exist: {0}", dropFolder);
Logger.Warn("Unable to Scan for New Downloads - folder Doesn't exist: [{0}]", dropFolder);
return;
}
@ -64,13 +63,13 @@ namespace NzbDrone.Core.Providers.Jobs
if (subfolderInfo.Name.StartsWith("_UNPACK_", StringComparison.CurrentCultureIgnoreCase))
{
Logger.Info("Folder [{0}] is still being unpacked. skipping.", subfolder);
Logger.Debug("Folder [{0}] is still being unpacked. skipping.", subfolder);
continue;
}
if (subfolderInfo.Name.StartsWith("_FAILED_", StringComparison.CurrentCultureIgnoreCase))
{
Logger.Info("Folder [{0}] is marked as failed. skipping.", subfolder);
Logger.Debug("Folder [{0}] is marked as failed. skipping.", subfolder);
continue;
}
@ -80,7 +79,7 @@ namespace NzbDrone.Core.Providers.Jobs
if (series == null)
{
Logger.Warn("Unable to Import new download, series is not being watched");
Logger.Warn("Unable to Import new download, series doesn't exist in database.");
return;
}
@ -96,8 +95,6 @@ namespace NzbDrone.Core.Providers.Jobs
Logger.ErrorException("An error has occurred while importing " + subfolder, e);
}
}
Logger.Debug("New Download Scan Job completed successfully");
}
}
}

View File

@ -69,10 +69,12 @@ namespace NzbDrone.Core.Providers.Jobs
}
catch (Exception e)
{
Logger.ErrorException("An error has occured while processing parse result items from " + episodeParseResult, e);
Logger.ErrorException("An error has occurred while processing parse result items from " + episodeParseResult, e);
}
}
notification.CurrentMessage = "RSS Sync Completed";
}
}
}