diff --git a/raspiwrite.py b/raspiwrite.py index 054f93a..c68a95d 100755 --- a/raspiwrite.py +++ b/raspiwrite.py @@ -11,8 +11,8 @@ # .kkx0NXk'...dNNxldK' # 'k...0o....,O...d: # ;o;'oM0olkWc.;oc -# .cOx....dOl. -# .x00k. +# .cOx....dOl. +# .x00k. #//////////////////////////// # * Raspberry Pi SD Writer @@ -22,17 +22,17 @@ # Copyright Matthew Jump 2012 # The following code is licenced under the Gnu Public Licence, please see gpl.txt for reference # This program is free software: you can redistribute it and/or modify -# it under the terms of the GNU General Public License as published by +# it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or -# (at your option) any later version. +# (at your option) any later version. # -# This program is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -# GNU General Public License for more details. +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. # -# You should have received a copy of the GNU General Public License -# along with this program. If not, see . +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . # VERSION 1.15 -MACOSX- (June 2012) BETA # * Fix to unzipping system - credit to alecthegeek @@ -84,8 +84,8 @@ def checkforUpdate(): ''' except urllib2.URLError, e: - print """There was an error in checking for an update: %r - """ % e + print """There was an error in checking for an update: %r + """ % e def grabRoot(distro): #Parses the raspberry pi downloads page for the links for the currently RasPiWrite supported distros @@ -103,8 +103,8 @@ def getZipUrl(url): #gets all the urls that end in .zip or .tar.bz2 (only two di htmlSource = urllib2.urlopen(url).read() linksList = re.findall('