diff --git a/raspiwrite.py b/raspiwrite.py
index 054f93a..c68a95d 100755
--- a/raspiwrite.py
+++ b/raspiwrite.py
@@ -11,8 +11,8 @@
# .kkx0NXk'...dNNxldK'
# 'k...0o....,O...d:
# ;o;'oM0olkWc.;oc
-# .cOx....dOl.
-# .x00k.
+# .cOx....dOl.
+# .x00k.
#////////////////////////////
# * Raspberry Pi SD Writer
@@ -22,17 +22,17 @@
# Copyright Matthew Jump 2012
# The following code is licenced under the Gnu Public Licence, please see gpl.txt for reference
# This program is free software: you can redistribute it and/or modify
-# it under the terms of the GNU General Public License as published by
+# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
-# (at your option) any later version.
+# (at your option) any later version.
#
-# This program is distributed in the hope that it will be useful,
-# but WITHOUT ANY WARRANTY; without even the implied warranty of
-# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-# GNU General Public License for more details.
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
#
-# You should have received a copy of the GNU General Public License
-# along with this program. If not, see .
+# You should have received a copy of the GNU General Public License
+# along with this program. If not, see .
# VERSION 1.15 -MACOSX- (June 2012) BETA
# * Fix to unzipping system - credit to alecthegeek
@@ -84,8 +84,8 @@ def checkforUpdate():
'''
except urllib2.URLError, e:
- print """There was an error in checking for an update: %r
- """ % e
+ print """There was an error in checking for an update: %r
+ """ % e
def grabRoot(distro): #Parses the raspberry pi downloads page for the links for the currently RasPiWrite supported distros
@@ -103,8 +103,8 @@ def getZipUrl(url): #gets all the urls that end in .zip or .tar.bz2 (only two di
htmlSource = urllib2.urlopen(url).read()
linksList = re.findall('