Browse code

.: remove trailing white spaces

blame tibor this one ;-)
```
find . -type f -not -name '*.png' -not -name '*.go' -not -name '*.md'
-not -name '*.tar' -not -name '*.pem' -not -path './vendor/*' -not -path
'./.git/*' -not -path '*/testdata/*' -not -path './docs/*images*' -not
-path '*/testfiles/*' -not -path './bundles/*' -not -path
'./docs/*static*/*' -not -path './docs/*article-img/*' -exec grep -HnEl
'[[:space:]]$' {} \; | xargs sed -iE 's/[[:space:]]*$//'
```

Signed-off-by: Vincent Batts <vbatts@redhat.com>

Vincent Batts authored on 2015/03/26 02:38:17
Showing 20 changed files
... ...
@@ -113,7 +113,7 @@ It is the responsibility of the subsystem maintainers to process patches affecti
113 113
 manner.
114 114
 
115 115
 * If the change affects areas of the code which are not part of a subsystem,
116
-or if subsystem maintainers are unable to reach a timely decision, it must be approved by 
116
+or if subsystem maintainers are unable to reach a timely decision, it must be approved by
117 117
 the core maintainers.
118 118
 
119 119
 * If the change affects the UI or public APIs, or if it represents a major change in architecture,
... ...
@@ -200,11 +200,11 @@ for each.
200 200
 			2-code-review = "requires more code changes"
201 201
 			1-design-review = "raises design concerns"
202 202
 			4-merge = "general case"
203
-			
203
+
204 204
 		# Docs approval
205 205
 		[Rules.review.docs-approval]
206 206
 			# Changes and additions to docs must be reviewed and approved (LGTM'd) by a minimum of two docs sub-project maintainers.
207
-			# If the docs change originates with a docs maintainer, only one additional LGTM is required (since we assume a docs maintainer approves of their own PR). 	
207
+			# If the docs change originates with a docs maintainer, only one additional LGTM is required (since we assume a docs maintainer approves of their own PR).
208 208
 
209 209
 		# Merge
210 210
 		[Rules.review.states.4-merge]
... ...
@@ -268,7 +268,7 @@ made through a pull request.
268 268
 
269 269
 	# The chief architect is responsible for the overall integrity of the technical architecture
270 270
 	# across all subsystems, and the consistency of APIs and UI.
271
-	# 
271
+	#
272 272
 	# Changes to UI, public APIs and overall architecture (for example a plugin system) must
273 273
 	# be approved by the chief architect.
274 274
 	"Chief Architect" = "shykes"
... ...
@@ -314,7 +314,7 @@ made through a pull request.
314 314
 			]
315 315
 
316 316
 	# The chief maintainer is responsible for all aspects of quality for the project including
317
-	# code reviews, usability, stability, security, performance, etc. 
317
+	# code reviews, usability, stability, security, performance, etc.
318 318
 	# The most important function of the chief maintainer is to lead by example. On the first
319 319
 	# day of a new maintainer, the best advice should be "follow the C.M.'s example and you'll
320 320
 	# be fine".
... ...
@@ -359,9 +359,9 @@ made through a pull request.
359 359
 	# has a dedicated group of maintainers, which are dedicated to that subsytem and responsible
360 360
 	# for its quality.
361 361
 	# This "cellular division" is the primary mechanism for scaling maintenance of the project as it grows.
362
-	# 
362
+	#
363 363
 	# The maintainers of each subsytem are responsible for:
364
-	# 
364
+	#
365 365
 	# 1. Exposing a clear road map for improving their subsystem.
366 366
 	# 2. Deliver prompt feedback and decisions on pull requests affecting their subsystem.
367 367
 	# 3. Be available to anyone with questions, bug reports, criticism etc.
... ...
@@ -371,9 +371,9 @@ made through a pull request.
371 371
 	#   road map of the project.
372 372
 	#
373 373
 	# #### How to review patches to your subsystem
374
-	# 
374
+	#
375 375
 	# Accepting pull requests:
376
-	# 
376
+	#
377 377
 	#   - If the pull request appears to be ready to merge, give it a `LGTM`, which
378 378
 	#     stands for "Looks Good To Me".
379 379
 	#   - If the pull request has some small problems that need to be changed, make
... ...
@@ -384,9 +384,9 @@ made through a pull request.
384 384
 	#   - If the PR only needs a few changes before being merged, any MAINTAINER can
385 385
 	#     make a replacement PR that incorporates the existing commits and fixes the
386 386
 	#     problems before a fast track merge.
387
-	# 
387
+	#
388 388
 	# Closing pull requests:
389
-	# 
389
+	#
390 390
 	#   - If a PR appears to be abandoned, after having attempted to contact the
391 391
 	#     original contributor, then a replacement PR may be made.  Once the
392 392
 	#     replacement PR is made, any contributor may close the original one.
... ...
@@ -584,12 +584,12 @@ made through a pull request.
584 584
 	Name = "Solomon Hykes"
585 585
 	Email = "solomon@docker.com"
586 586
 	GitHub = "shykes"
587
-    
587
+
588 588
 	[people.spf13]
589 589
 	Name = "Steve Francia"
590 590
 	Email = "steve.francia@gmail.com"
591 591
 	GitHub = "spf13"
592
-	
592
+
593 593
 	[people.sven]
594 594
 	Name = "Sven Dowideit"
595 595
 	Email = "SvenDowideit@home.org.au"
... ...
@@ -10,9 +10,9 @@ The following is courtesy of our legal counsel:
10 10
 
11 11
 
12 12
 Use and transfer of Docker may be subject to certain restrictions by the
13
-United States and other governments.  
13
+United States and other governments.
14 14
 It is your responsibility to ensure that your use and/or transfer does not
15
-violate applicable laws. 
15
+violate applicable laws.
16 16
 
17 17
 For more information, please see http://www.bis.doc.gov
18 18
 
... ...
@@ -15,7 +15,7 @@ hello\\                  |     hello\
15 15
 'hello\'                 |     hello\
16 16
 "''"                     |     ''
17 17
 $.                       |     $.
18
-$1                       |     
18
+$1                       |
19 19
 he$1x                    |     hex
20 20
 he$.x                    |     he$.x
21 21
 he$pwd.                  |     he.
... ...
@@ -22,7 +22,7 @@
22 22
 # must have access to the socket for the completions to function correctly
23 23
 #
24 24
 # Note for developers:
25
-# Please arrange options sorted alphabetically by long name with the short 
25
+# Please arrange options sorted alphabetically by long name with the short
26 26
 # options immediately following their corresponding long form.
27 27
 # This order should be applied to lists, alternatives and code blocks.
28 28
 
... ...
@@ -257,8 +257,8 @@ _docker_build() {
257 257
 			;;
258 258
 		--file|-f)
259 259
 			_filedir
260
-			return	
261
-			;;	
260
+			return
261
+			;;
262 262
 	esac
263 263
 
264 264
 	case "$cur" in
... ...
@@ -3,7 +3,7 @@
3 3
 # AUTHOR:         Jessica Frazelle <jess@docker.com>
4 4
 # COMMENTS:
5 5
 #   This file describes how to build a gparted container with all
6
-#   dependencies installed. It uses native X11 unix socket. 
6
+#   dependencies installed. It uses native X11 unix socket.
7 7
 #   Tested on Debian Jessie
8 8
 # USAGE:
9 9
 #   # Download gparted Dockerfile
... ...
@@ -41,39 +41,39 @@ while [ $# -gt 0 ]; do
41 41
 	[ "$imageId" != "$tag" ] || imageId=
42 42
 	[ "$tag" != "$imageTag" ] || tag='latest'
43 43
 	tag="${tag%@*}"
44
-	
44
+
45 45
 	token="$(curl -sSL -o /dev/null -D- -H 'X-Docker-Token: true' "https://index.docker.io/v1/repositories/$image/images" | tr -d '\r' | awk -F ': *' '$1 == "X-Docker-Token" { print $2 }')"
46
-	
46
+
47 47
 	if [ -z "$imageId" ]; then
48 48
 		imageId="$(curl -sSL -H "Authorization: Token $token" "https://registry-1.docker.io/v1/repositories/$image/tags/$tag")"
49 49
 		imageId="${imageId//\"/}"
50 50
 	fi
51
-	
51
+
52 52
 	ancestryJson="$(curl -sSL -H "Authorization: Token $token" "https://registry-1.docker.io/v1/images/$imageId/ancestry")"
53 53
 	if [ "${ancestryJson:0:1}" != '[' ]; then
54 54
 		echo >&2 "error: /v1/images/$imageId/ancestry returned something unexpected:"
55 55
 		echo >&2 "  $ancestryJson"
56 56
 		exit 1
57 57
 	fi
58
-	
58
+
59 59
 	IFS=','
60 60
 	ancestry=( ${ancestryJson//[\[\] \"]/} )
61 61
 	unset IFS
62
-	
62
+
63 63
 	if [ -s "$dir/tags-$image.tmp" ]; then
64 64
 		echo -n ', ' >> "$dir/tags-$image.tmp"
65 65
 	else
66 66
 		images=( "${images[@]}" "$image" )
67 67
 	fi
68 68
 	echo -n '"'"$tag"'": "'"$imageId"'"' >> "$dir/tags-$image.tmp"
69
-	
69
+
70 70
 	echo "Downloading '$imageTag' (${#ancestry[@]} layers)..."
71 71
 	for imageId in "${ancestry[@]}"; do
72 72
 		mkdir -p "$dir/$imageId"
73 73
 		echo '1.0' > "$dir/$imageId/VERSION"
74
-		
74
+
75 75
 		curl -sSL -H "Authorization: Token $token" "https://registry-1.docker.io/v1/images/$imageId/json" -o "$dir/$imageId/json"
76
-		
76
+
77 77
 		# TODO figure out why "-C -" doesn't work here
78 78
 		# "curl: (33) HTTP server doesn't seem to support byte ranges. Cannot resume."
79 79
 		# "HTTP/1.1 416 Requested Range Not Satisfiable"
... ...
@@ -10,11 +10,11 @@ cat <<-EOF
10 10
 	Description=$desc
11 11
 	Author=$auth
12 12
 	After=docker.service
13
-	
13
+
14 14
 	[Service]
15 15
 	ExecStart=/usr/bin/docker start -a $cid
16 16
 	ExecStop=/usr/bin/docker stop -t 2 $cid
17
-	
17
+
18 18
 	[Install]
19 19
 	WantedBy=local.target
20 20
 EOF
... ...
@@ -1,5 +1,5 @@
1 1
 # /etc/sysconfig/docker
2
-# 
2
+#
3 3
 # Other arguments to pass to the docker daemon process
4 4
 # These will be parsed by the sysv initscript and appended
5 5
 # to the arguments list passed to docker -d
... ...
@@ -14,9 +14,9 @@ justTar=
14 14
 
15 15
 usage() {
16 16
 	echo >&2
17
-	
17
+
18 18
 	echo >&2 "usage: $0 [options] repo suite [mirror]"
19
-	
19
+
20 20
 	echo >&2
21 21
 	echo >&2 'options: (not recommended)'
22 22
 	echo >&2 "  -p set an http_proxy for debootstrap"
... ...
@@ -26,20 +26,20 @@ usage() {
26 26
 	echo >&2 "  -s # skip version detection and tagging (ie, precise also tagged as 12.04)"
27 27
 	echo >&2 "     # note that this will also skip adding universe and/or security/updates to sources.list"
28 28
 	echo >&2 "  -t # just create a tarball, especially for dockerbrew (uses repo as tarball name)"
29
-	
29
+
30 30
 	echo >&2
31 31
 	echo >&2 "   ie: $0 username/debian squeeze"
32 32
 	echo >&2 "       $0 username/debian squeeze http://ftp.uk.debian.org/debian/"
33
-	
33
+
34 34
 	echo >&2
35 35
 	echo >&2 "   ie: $0 username/ubuntu precise"
36 36
 	echo >&2 "       $0 username/ubuntu precise http://mirrors.melbourne.co.uk/ubuntu/"
37
-	
37
+
38 38
 	echo >&2
39 39
 	echo >&2 "   ie: $0 -t precise.tar.bz2 precise"
40 40
 	echo >&2 "       $0 -t wheezy.tgz wheezy"
41 41
 	echo >&2 "       $0 -t wheezy-uk.tar.xz wheezy http://ftp.uk.debian.org/debian/"
42
-	
42
+
43 43
 	echo >&2
44 44
 }
45 45
 
... ...
@@ -145,10 +145,10 @@ if [ -z "$strictDebootstrap" ]; then
145 145
 	sudo chroot . dpkg-divert --local --rename --add /sbin/initctl
146 146
 	sudo ln -sf /bin/true sbin/initctl
147 147
 	# see https://github.com/docker/docker/issues/446#issuecomment-16953173
148
-	
148
+
149 149
 	# shrink the image, since apt makes us fat (wheezy: ~157.5MB vs ~120MB)
150 150
 	sudo chroot . apt-get clean
151
-	
151
+
152 152
 	if strings usr/bin/dpkg | grep -q unsafe-io; then
153 153
 		# while we're at it, apt is unnecessarily slow inside containers
154 154
 		#  this forces dpkg not to call sync() after package extraction and speeds up install
... ...
@@ -159,7 +159,7 @@ if [ -z "$strictDebootstrap" ]; then
159 159
 		# (see http://bugs.debian.org/cgi-bin/bugreport.cgi?bug=584254#82),
160 160
 		# and ubuntu lucid/10.04 only has 1.15.5.6
161 161
 	fi
162
-	
162
+
163 163
 	# we want to effectively run "apt-get clean" after every install to keep images small (see output of "apt-get clean -s" for context)
164 164
 	{
165 165
 		aptGetClean='"rm -f /var/cache/apt/archives/*.deb /var/cache/apt/archives/partial/*.deb /var/cache/apt/*.bin || true";'
... ...
@@ -167,17 +167,17 @@ if [ -z "$strictDebootstrap" ]; then
167 167
 		echo "APT::Update::Post-Invoke { ${aptGetClean} };"
168 168
 		echo 'Dir::Cache::pkgcache ""; Dir::Cache::srcpkgcache "";'
169 169
 	} | sudo tee etc/apt/apt.conf.d/no-cache > /dev/null
170
-	
170
+
171 171
 	# and remove the translations, too
172 172
 	echo 'Acquire::Languages "none";' | sudo tee etc/apt/apt.conf.d/no-languages > /dev/null
173
-	
173
+
174 174
 	# helpful undo lines for each the above tweaks (for lack of a better home to keep track of them):
175 175
 	#  rm /usr/sbin/policy-rc.d
176 176
 	#  rm /sbin/initctl; dpkg-divert --rename --remove /sbin/initctl
177 177
 	#  rm /etc/dpkg/dpkg.cfg.d/02apt-speedup
178 178
 	#  rm /etc/apt/apt.conf.d/no-cache
179 179
 	#  rm /etc/apt/apt.conf.d/no-languages
180
-	
180
+
181 181
 	if [ -z "$skipDetection" ]; then
182 182
 		# see also rudimentary platform detection in hack/install.sh
183 183
 		lsbDist=''
... ...
@@ -187,14 +187,14 @@ if [ -z "$strictDebootstrap" ]; then
187 187
 		if [ -z "$lsbDist" ] && [ -r etc/debian_version ]; then
188 188
 			lsbDist='Debian'
189 189
 		fi
190
-		
190
+
191 191
 		case "$lsbDist" in
192 192
 			Debian)
193 193
 				# add the updates and security repositories
194 194
 				if [ "$suite" != "$debianUnstable" -a "$suite" != 'unstable' ]; then
195 195
 					# ${suite}-updates only applies to non-unstable
196 196
 					sudo sed -i "p; s/ $suite main$/ ${suite}-updates main/" etc/apt/sources.list
197
-					
197
+
198 198
 					# same for security updates
199 199
 					echo "deb http://security.debian.org/ $suite/updates main" | sudo tee -a etc/apt/sources.list > /dev/null
200 200
 				fi
... ...
@@ -220,7 +220,7 @@ if [ -z "$strictDebootstrap" ]; then
220 220
 				;;
221 221
 		esac
222 222
 	fi
223
-	
223
+
224 224
 	# make sure our packages lists are as up to date as we can get them
225 225
 	sudo chroot . apt-get update
226 226
 	sudo chroot . apt-get dist-upgrade -y
... ...
@@ -229,23 +229,23 @@ fi
229 229
 if [ "$justTar" ]; then
230 230
 	# create the tarball file so it has the right permissions (ie, not root)
231 231
 	touch "$repo"
232
-	
232
+
233 233
 	# fill the tarball
234 234
 	sudo tar --numeric-owner -caf "$repo" .
235 235
 else
236 236
 	# create the image (and tag $repo:$suite)
237 237
 	sudo tar --numeric-owner -c . | $docker import - $repo:$suite
238
-	
238
+
239 239
 	# test the image
240 240
 	$docker run -i -t $repo:$suite echo success
241
-	
241
+
242 242
 	if [ -z "$skipDetection" ]; then
243 243
 		case "$lsbDist" in
244 244
 			Debian)
245 245
 				if [ "$suite" = "$debianStable" -o "$suite" = 'stable' ] && [ -r etc/debian_version ]; then
246 246
 					# tag latest
247 247
 					$docker tag $repo:$suite $repo:latest
248
-					
248
+
249 249
 					if [ -r etc/debian_version ]; then
250 250
 						# tag the specific debian release version (which is only reasonable to tag on debian stable)
251 251
 						ver=$(cat etc/debian_version)
... ...
@@ -19,7 +19,7 @@ shift
19 19
 chrootPath="$(type -P chroot)"
20 20
 rootfs_chroot() {
21 21
 	# "chroot" doesn't set PATH, so we need to set it explicitly to something our new debootstrap chroot can use appropriately!
22
-	
22
+
23 23
 	# set PATH and chroot away!
24 24
 	PATH='/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin' \
25 25
 		"$chrootPath" "$rootfsDir" "$@"
... ...
@@ -220,13 +220,13 @@ fi
220 220
 
221 221
 (
222 222
 	set -x
223
-	
223
+
224 224
 	# make sure we're fully up-to-date
225 225
 	rootfs_chroot sh -xc 'apt-get update && apt-get dist-upgrade -y'
226
-	
226
+
227 227
 	# delete all the apt list files since they're big and get stale quickly
228 228
 	rm -rf "$rootfsDir/var/lib/apt/lists"/*
229 229
 	# this forces "apt-get update" in dependent images, which is also good
230
-	
230
+
231 231
 	mkdir "$rootfsDir/var/lib/apt/lists/partial" # Lucid... "E: Lists directory /var/lib/apt/lists/partial is missing."
232 232
 )
... ...
@@ -29,41 +29,41 @@ function template() {
29 29
 # this should always match the template from CONTRIBUTING.md
30 30
 	cat <<- EOM
31 31
 	Description of problem:
32
-	
33
-	
32
+
33
+
34 34
 	\`docker version\`:
35 35
 	`${DOCKER_COMMAND} -D version`
36
-	
37
-	
36
+
37
+
38 38
 	\`docker info\`:
39 39
 	`${DOCKER_COMMAND} -D info`
40
-	
41
-	
40
+
41
+
42 42
 	\`uname -a\`:
43 43
 	`uname -a`
44
-	
45
-	
44
+
45
+
46 46
 	Environment details (AWS, VirtualBox, physical, etc.):
47
-	
48
-	
47
+
48
+
49 49
 	How reproducible:
50
-	
51
-	
50
+
51
+
52 52
 	Steps to Reproduce:
53 53
 	1.
54 54
 	2.
55 55
 	3.
56
-	
57
-	
56
+
57
+
58 58
 	Actual Results:
59
-	
60
-	
59
+
60
+
61 61
 	Expected Results:
62
-	
63
-	
62
+
63
+
64 64
 	Additional info:
65
-	
66
-	
65
+
66
+
67 67
 	EOM
68 68
 }
69 69
 
... ...
@@ -81,7 +81,7 @@ echo -ne "Do you use \`sudo\` to call docker? [y|N]: "
81 81
 read -r -n 1 use_sudo
82 82
 echo ""
83 83
 
84
-if [ "x${use_sudo}" = "xy" -o "x${use_sudo}" = "xY" ]; then 
84
+if [ "x${use_sudo}" = "xy" -o "x${use_sudo}" = "xY" ]; then
85 85
 	export DOCKER_COMMAND="sudo ${DOCKER}"
86 86
 fi
87 87
 
... ...
@@ -18,12 +18,12 @@
18 18
 				<key>0</key>
19 19
 				<dict>
20 20
 					<key>name</key>
21
-					<string>keyword.control.dockerfile</string>					
21
+					<string>keyword.control.dockerfile</string>
22 22
 				</dict>
23 23
 				<key>1</key>
24 24
 				<dict>
25 25
 					<key>name</key>
26
-					<string>keyword.other.special-method.dockerfile</string>					
26
+					<string>keyword.other.special-method.dockerfile</string>
27 27
 				</dict>
28 28
 			</dict>
29 29
 		</dict>
... ...
@@ -35,12 +35,12 @@
35 35
 				<key>0</key>
36 36
 				<dict>
37 37
 					<key>name</key>
38
-					<string>keyword.operator.dockerfile</string>					
38
+					<string>keyword.operator.dockerfile</string>
39 39
 				</dict>
40 40
 				<key>1</key>
41 41
 				<dict>
42 42
 					<key>name</key>
43
-					<string>keyword.other.special-method.dockerfile</string>					
43
+					<string>keyword.other.special-method.dockerfile</string>
44 44
 				</dict>
45 45
 			</dict>
46 46
 		</dict>
... ...
@@ -176,15 +176,15 @@ pages:
176 176
 # Project:
177 177
 - ['project/index.md', '**HIDDEN**']
178 178
 - ['project/who-written-for.md', 'Contributor Guide', 'README first']
179
-- ['project/software-required.md', 'Contributor Guide', 'Get required software'] 
180
-- ['project/set-up-git.md', 'Contributor Guide', 'Configure Git for contributing'] 
181
-- ['project/set-up-dev-env.md', 'Contributor Guide', 'Work with a development container'] 
179
+- ['project/software-required.md', 'Contributor Guide', 'Get required software']
180
+- ['project/set-up-git.md', 'Contributor Guide', 'Configure Git for contributing']
181
+- ['project/set-up-dev-env.md', 'Contributor Guide', 'Work with a development container']
182 182
 - ['project/test-and-docs.md', 'Contributor Guide', 'Run tests and test documentation']
183 183
 - ['project/make-a-contribution.md', 'Contributor Guide', 'Understand contribution workflow']
184
-- ['project/find-an-issue.md', 'Contributor Guide', 'Find an issue'] 
185
-- ['project/work-issue.md', 'Contributor Guide', 'Work on an issue'] 
186
-- ['project/create-pr.md', 'Contributor Guide', 'Create a pull request'] 
187
-- ['project/review-pr.md', 'Contributor Guide', 'Participate in the PR review'] 
184
+- ['project/find-an-issue.md', 'Contributor Guide', 'Find an issue']
185
+- ['project/work-issue.md', 'Contributor Guide', 'Work on an issue']
186
+- ['project/create-pr.md', 'Contributor Guide', 'Create a pull request']
187
+- ['project/review-pr.md', 'Contributor Guide', 'Participate in the PR review']
188 188
 - ['project/advanced-contributing.md', 'Contributor Guide', 'Advanced contributing']
189 189
 - ['project/get-help.md', 'Contributor Guide', 'Where to get help']
190 190
 - ['project/coding-style.md', 'Contributor Guide', 'Coding style guide']
... ...
@@ -5,7 +5,7 @@ set -o pipefail
5 5
 
6 6
 usage() {
7 7
 	cat >&2 <<'EOF'
8
-To publish the Docker documentation you need to set your access_key and secret_key in the docs/awsconfig file 
8
+To publish the Docker documentation you need to set your access_key and secret_key in the docs/awsconfig file
9 9
 (with the keys in a [profile $AWS_S3_BUCKET] section - so you can have more than one set of keys in your file)
10 10
 and set the AWS_S3_BUCKET env var to the name of your bucket.
11 11
 
... ...
@@ -6,7 +6,7 @@ FROM ubuntu
6 6
 MAINTAINER SvenDowideit@docker.com
7 7
 
8 8
 # Add the PostgreSQL PGP key to verify their Debian packages.
9
-# It should be the same key as https://www.postgresql.org/media/keys/ACCC4CF8.asc 
9
+# It should be the same key as https://www.postgresql.org/media/keys/ACCC4CF8.asc
10 10
 RUN apt-key adv --keyserver hkp://p80.pool.sks-keyservers.net:80 --recv-keys B97B0AFCAA1A47F044F244A07FCC7D46ACCC4CF8
11 11
 
12 12
 # Add PostgreSQL's repository. It contains the most recent stable release
... ...
@@ -33,7 +33,7 @@ RUN    /etc/init.d/postgresql start &&\
33 33
     createdb -O docker docker
34 34
 
35 35
 # Adjust PostgreSQL configuration so that remote connections to the
36
-# database are possible. 
36
+# database are possible.
37 37
 RUN echo "host all  all    0.0.0.0/0  md5" >> /etc/postgresql/9.3/main/pg_hba.conf
38 38
 
39 39
 # And add ``listen_addresses`` to ``/etc/postgresql/9.3/main/postgresql.conf``
... ...
@@ -3,23 +3,23 @@
3 3
 if [ -z "$VALIDATE_UPSTREAM" ]; then
4 4
 	# this is kind of an expensive check, so let's not do this twice if we
5 5
 	# are running more than one validate bundlescript
6
-	
6
+
7 7
 	VALIDATE_REPO='https://github.com/docker/docker.git'
8 8
 	VALIDATE_BRANCH='master'
9
-	
9
+
10 10
 	if [ "$TRAVIS" = 'true' -a "$TRAVIS_PULL_REQUEST" != 'false' ]; then
11 11
 		VALIDATE_REPO="https://github.com/${TRAVIS_REPO_SLUG}.git"
12 12
 		VALIDATE_BRANCH="${TRAVIS_BRANCH}"
13 13
 	fi
14
-	
14
+
15 15
 	VALIDATE_HEAD="$(git rev-parse --verify HEAD)"
16
-	
16
+
17 17
 	git fetch -q "$VALIDATE_REPO" "refs/heads/$VALIDATE_BRANCH"
18 18
 	VALIDATE_UPSTREAM="$(git rev-parse --verify FETCH_HEAD)"
19
-	
19
+
20 20
 	VALIDATE_COMMIT_LOG="$VALIDATE_UPSTREAM..$VALIDATE_HEAD"
21 21
 	VALIDATE_COMMIT_DIFF="$VALIDATE_UPSTREAM...$VALIDATE_HEAD"
22
-	
22
+
23 23
 	validate_diff() {
24 24
 		if [ "$VALIDATE_UPSTREAM" != "$VALIDATE_HEAD" ]; then
25 25
 			git diff "$VALIDATE_COMMIT_DIFF" "$@"
... ...
@@ -5,7 +5,7 @@ DEST=$1
5 5
 
6 6
 if [ -z "$DOCKER_CLIENTONLY" ]; then
7 7
 	source "$(dirname "$BASH_SOURCE")/.dockerinit"
8
-	
8
+
9 9
 	hash_files "$DEST/dockerinit-$VERSION"
10 10
 else
11 11
 	# DOCKER_CLIENTONLY must be truthy, so we don't need to bother with dockerinit :)
... ...
@@ -5,7 +5,7 @@ DEST=$1
5 5
 
6 6
 if [ -z "$DOCKER_CLIENTONLY" ]; then
7 7
 	source "$(dirname "$BASH_SOURCE")/.dockerinit-gccgo"
8
-	
8
+
9 9
 	hash_files "$DEST/dockerinit-$VERSION"
10 10
 else
11 11
 	# DOCKER_CLIENTONLY must be truthy, so we don't need to bother with dockerinit :)
... ...
@@ -18,17 +18,17 @@ for d in "$CROSS/"*/*; do
18 18
 	BINARY_FULLNAME="$BINARY_NAME$BINARY_EXTENSION"
19 19
 	mkdir -p "$DEST/$GOOS/$GOARCH"
20 20
 	TGZ="$DEST/$GOOS/$GOARCH/$BINARY_NAME.tgz"
21
-	
21
+
22 22
 	mkdir -p "$DEST/build"
23
-	
23
+
24 24
 	mkdir -p "$DEST/build/usr/local/bin"
25 25
 	cp -L "$d/$BINARY_FULLNAME" "$DEST/build/usr/local/bin/docker$BINARY_EXTENSION"
26
-	
26
+
27 27
 	tar --numeric-owner --owner 0 -C "$DEST/build" -czf "$TGZ" usr
28
-	
28
+
29 29
 	hash_files "$TGZ"
30
-	
30
+
31 31
 	rm -rf "$DEST/build"
32
-	
32
+
33 33
 	echo "Created tgz: $TGZ"
34 34
 done
... ...
@@ -11,17 +11,17 @@ clone() {
11 11
 	vcs=$1
12 12
 	pkg=$2
13 13
 	rev=$3
14
-	
14
+
15 15
 	pkg_url=https://$pkg
16 16
 	target_dir=src/$pkg
17
-	
17
+
18 18
 	echo -n "$pkg @ $rev: "
19
-	
19
+
20 20
 	if [ -d $target_dir ]; then
21 21
 		echo -n 'rm old, '
22 22
 		rm -fr $target_dir
23 23
 	fi
24
-	
24
+
25 25
 	echo -n 'clone, '
26 26
 	case $vcs in
27 27
 		git)
... ...
@@ -32,10 +32,10 @@ clone() {
32 32
 			hg clone --quiet --updaterev $rev $pkg_url $target_dir
33 33
 			;;
34 34
 	esac
35
-	
35
+
36 36
 	echo -n 'rm VCS, '
37 37
 	( cd $target_dir && rm -rf .{git,hg} )
38
-	
38
+
39 39
 	echo done
40 40
 }
41 41