]>
Commit | Line | Data |
---|---|---|
24fd1fe6 IB |
1 | # This file originates from node2nix |
2 | ||
3 | {stdenv, nodejs, python2, utillinux, libtool, runCommand, writeTextFile}: | |
4 | ||
5 | let | |
6 | python = if nodejs ? python then nodejs.python else python2; | |
7 | ||
8 | # Create a tar wrapper that filters all the 'Ignoring unknown extended header keyword' noise | |
9 | tarWrapper = runCommand "tarWrapper" {} '' | |
10 | mkdir -p $out/bin | |
11 | ||
12 | cat > $out/bin/tar <<EOF | |
13 | #! ${stdenv.shell} -e | |
14 | $(type -p tar) "\$@" --warning=no-unknown-keyword | |
15 | EOF | |
16 | ||
17 | chmod +x $out/bin/tar | |
18 | ''; | |
19 | ||
20 | # Function that generates a TGZ file from a NPM project | |
21 | buildNodeSourceDist = | |
22 | { name, version, src, ... }: | |
23 | ||
24 | stdenv.mkDerivation { | |
25 | name = "node-tarball-${name}-${version}"; | |
26 | inherit src; | |
27 | buildInputs = [ nodejs ]; | |
28 | buildPhase = '' | |
29 | export HOME=$TMPDIR | |
30 | tgzFile=$(npm pack | tail -n 1) # Hooks to the pack command will add output (https://docs.npmjs.com/misc/scripts) | |
31 | ''; | |
32 | installPhase = '' | |
33 | mkdir -p $out/tarballs | |
34 | mv $tgzFile $out/tarballs | |
35 | mkdir -p $out/nix-support | |
36 | echo "file source-dist $out/tarballs/$tgzFile" >> $out/nix-support/hydra-build-products | |
37 | ''; | |
38 | }; | |
39 | ||
40 | includeDependencies = {dependencies}: | |
41 | stdenv.lib.optionalString (dependencies != []) | |
42 | (stdenv.lib.concatMapStrings (dependency: | |
43 | '' | |
44 | # Bundle the dependencies of the package | |
45 | mkdir -p node_modules | |
46 | cd node_modules | |
47 | ||
48 | # Only include dependencies if they don't exist. They may also be bundled in the package. | |
49 | if [ ! -e "${dependency.name}" ] | |
50 | then | |
51 | ${composePackage dependency} | |
52 | fi | |
53 | ||
54 | cd .. | |
55 | '' | |
56 | ) dependencies); | |
57 | ||
58 | # Recursively composes the dependencies of a package | |
59 | composePackage = { name, packageName, src, dependencies ? [], ... }@args: | |
60 | '' | |
61 | DIR=$(pwd) | |
62 | cd $TMPDIR | |
63 | ||
64 | unpackFile ${src} | |
65 | ||
66 | # Make the base dir in which the target dependency resides first | |
67 | mkdir -p "$(dirname "$DIR/${packageName}")" | |
68 | ||
69 | if [ -f "${src}" ] | |
70 | then | |
71 | # Figure out what directory has been unpacked | |
72 | packageDir="$(find . -maxdepth 1 -type d | tail -1)" | |
73 | ||
74 | # Restore write permissions to make building work | |
75 | find "$packageDir" -type d -print0 | xargs -0 chmod u+x | |
76 | chmod -R u+w "$packageDir" | |
77 | ||
78 | # Move the extracted tarball into the output folder | |
79 | mv "$packageDir" "$DIR/${packageName}" | |
80 | elif [ -d "${src}" ] | |
81 | then | |
82 | # Get a stripped name (without hash) of the source directory. | |
83 | # On old nixpkgs it's already set internally. | |
84 | if [ -z "$strippedName" ] | |
85 | then | |
86 | strippedName="$(stripHash ${src})" | |
87 | fi | |
88 | ||
89 | # Restore write permissions to make building work | |
90 | chmod -R u+w "$strippedName" | |
91 | ||
92 | # Move the extracted directory into the output folder | |
93 | mv "$strippedName" "$DIR/${packageName}" | |
94 | fi | |
95 | ||
96 | # Unset the stripped name to not confuse the next unpack step | |
97 | unset strippedName | |
98 | ||
99 | # Include the dependencies of the package | |
100 | cd "$DIR/${packageName}" | |
101 | ${includeDependencies { inherit dependencies; }} | |
102 | cd .. | |
103 | ${stdenv.lib.optionalString (builtins.substring 0 1 packageName == "@") "cd .."} | |
104 | ''; | |
105 | ||
106 | pinpointDependencies = {dependencies, production}: | |
107 | let | |
108 | pinpointDependenciesFromPackageJSON = writeTextFile { | |
109 | name = "pinpointDependencies.js"; | |
110 | text = '' | |
111 | var fs = require('fs'); | |
112 | var path = require('path'); | |
113 | ||
114 | function resolveDependencyVersion(location, name) { | |
115 | if(location == process.env['NIX_STORE']) { | |
116 | return null; | |
117 | } else { | |
118 | var dependencyPackageJSON = path.join(location, "node_modules", name, "package.json"); | |
119 | ||
120 | if(fs.existsSync(dependencyPackageJSON)) { | |
121 | var dependencyPackageObj = JSON.parse(fs.readFileSync(dependencyPackageJSON)); | |
122 | ||
123 | if(dependencyPackageObj.name == name) { | |
124 | return dependencyPackageObj.version; | |
125 | } | |
126 | } else { | |
127 | return resolveDependencyVersion(path.resolve(location, ".."), name); | |
128 | } | |
129 | } | |
130 | } | |
131 | ||
132 | function replaceDependencies(dependencies) { | |
133 | if(typeof dependencies == "object" && dependencies !== null) { | |
134 | for(var dependency in dependencies) { | |
135 | var resolvedVersion = resolveDependencyVersion(process.cwd(), dependency); | |
136 | ||
137 | if(resolvedVersion === null) { | |
138 | process.stderr.write("WARNING: cannot pinpoint dependency: "+dependency+", context: "+process.cwd()+"\n"); | |
139 | } else { | |
140 | dependencies[dependency] = resolvedVersion; | |
141 | } | |
142 | } | |
143 | } | |
144 | } | |
145 | ||
146 | /* Read the package.json configuration */ | |
147 | var packageObj = JSON.parse(fs.readFileSync('./package.json')); | |
148 | ||
149 | /* Pinpoint all dependencies */ | |
150 | replaceDependencies(packageObj.dependencies); | |
151 | if(process.argv[2] == "development") { | |
152 | replaceDependencies(packageObj.devDependencies); | |
153 | } | |
154 | replaceDependencies(packageObj.optionalDependencies); | |
155 | ||
156 | /* Write the fixed package.json file */ | |
157 | fs.writeFileSync("package.json", JSON.stringify(packageObj, null, 2)); | |
158 | ''; | |
159 | }; | |
160 | in | |
161 | '' | |
162 | node ${pinpointDependenciesFromPackageJSON} ${if production then "production" else "development"} | |
163 | ||
164 | ${stdenv.lib.optionalString (dependencies != []) | |
165 | '' | |
166 | if [ -d node_modules ] | |
167 | then | |
168 | cd node_modules | |
169 | ${stdenv.lib.concatMapStrings (dependency: pinpointDependenciesOfPackage dependency) dependencies} | |
170 | cd .. | |
171 | fi | |
172 | ''} | |
173 | ''; | |
174 | ||
175 | # Recursively traverses all dependencies of a package and pinpoints all | |
176 | # dependencies in the package.json file to the versions that are actually | |
177 | # being used. | |
178 | ||
179 | pinpointDependenciesOfPackage = { packageName, dependencies ? [], production ? true, ... }@args: | |
180 | '' | |
181 | if [ -d "${packageName}" ] | |
182 | then | |
183 | cd "${packageName}" | |
184 | ${pinpointDependencies { inherit dependencies production; }} | |
185 | cd .. | |
186 | ${stdenv.lib.optionalString (builtins.substring 0 1 packageName == "@") "cd .."} | |
187 | fi | |
188 | ''; | |
189 | ||
190 | # Extract the Node.js source code which is used to compile packages with | |
191 | # native bindings | |
192 | nodeSources = runCommand "node-sources" {} '' | |
193 | tar --no-same-owner --no-same-permissions -xf ${nodejs.src} | |
194 | mv node-* $out | |
195 | ''; | |
196 | ||
197 | # Script that adds _integrity fields to all package.json files to prevent NPM from consulting the cache (that is empty) | |
198 | addIntegrityFieldsScript = writeTextFile { | |
199 | name = "addintegrityfields.js"; | |
200 | text = '' | |
201 | var fs = require('fs'); | |
202 | var path = require('path'); | |
203 | ||
204 | function augmentDependencies(baseDir, dependencies) { | |
205 | for(var dependencyName in dependencies) { | |
206 | var dependency = dependencies[dependencyName]; | |
207 | ||
208 | // Open package.json and augment metadata fields | |
209 | var packageJSONDir = path.join(baseDir, "node_modules", dependencyName); | |
210 | var packageJSONPath = path.join(packageJSONDir, "package.json"); | |
211 | ||
212 | if(fs.existsSync(packageJSONPath)) { // Only augment packages that exist. Sometimes we may have production installs in which development dependencies can be ignored | |
213 | console.log("Adding metadata fields to: "+packageJSONPath); | |
214 | var packageObj = JSON.parse(fs.readFileSync(packageJSONPath)); | |
215 | ||
216 | if(dependency.integrity) { | |
217 | packageObj["_integrity"] = dependency.integrity; | |
218 | } else { | |
219 | packageObj["_integrity"] = "sha1-000000000000000000000000000="; // When no _integrity string has been provided (e.g. by Git dependencies), add a dummy one. It does not seem to harm and it bypasses downloads. | |
220 | } | |
221 | ||
222 | packageObj["_resolved"] = dependency.version; // Set the resolved version to the version identifier. This prevents NPM from cloning Git repositories. | |
223 | fs.writeFileSync(packageJSONPath, JSON.stringify(packageObj, null, 2)); | |
224 | } | |
225 | ||
226 | // Augment transitive dependencies | |
227 | if(dependency.dependencies !== undefined) { | |
228 | augmentDependencies(packageJSONDir, dependency.dependencies); | |
229 | } | |
230 | } | |
231 | } | |
232 | ||
233 | if(fs.existsSync("./package-lock.json")) { | |
234 | var packageLock = JSON.parse(fs.readFileSync("./package-lock.json")); | |
235 | ||
236 | if(packageLock.lockfileVersion !== 1) { | |
237 | process.stderr.write("Sorry, I only understand lock file version 1!\n"); | |
238 | process.exit(1); | |
239 | } | |
240 | ||
241 | if(packageLock.dependencies !== undefined) { | |
242 | augmentDependencies(".", packageLock.dependencies); | |
243 | } | |
244 | } | |
245 | ''; | |
246 | }; | |
247 | ||
248 | # Reconstructs a package-lock file from the node_modules/ folder structure and package.json files with dummy sha1 hashes | |
249 | reconstructPackageLock = writeTextFile { | |
250 | name = "addintegrityfields.js"; | |
251 | text = '' | |
252 | var fs = require('fs'); | |
253 | var path = require('path'); | |
254 | ||
255 | var packageObj = JSON.parse(fs.readFileSync("package.json")); | |
256 | ||
257 | var lockObj = { | |
258 | name: packageObj.name, | |
259 | version: packageObj.version, | |
260 | lockfileVersion: 1, | |
261 | requires: true, | |
262 | dependencies: {} | |
263 | }; | |
264 | ||
265 | function augmentPackageJSON(filePath, dependencies) { | |
266 | var packageJSON = path.join(filePath, "package.json"); | |
267 | if(fs.existsSync(packageJSON)) { | |
268 | var packageObj = JSON.parse(fs.readFileSync(packageJSON)); | |
269 | dependencies[packageObj.name] = { | |
270 | version: packageObj.version, | |
271 | integrity: "sha1-000000000000000000000000000=", | |
272 | dependencies: {} | |
273 | }; | |
274 | processDependencies(path.join(filePath, "node_modules"), dependencies[packageObj.name].dependencies); | |
275 | } | |
276 | } | |
277 | ||
278 | function processDependencies(dir, dependencies) { | |
279 | if(fs.existsSync(dir)) { | |
280 | var files = fs.readdirSync(dir); | |
281 | ||
282 | files.forEach(function(entry) { | |
283 | var filePath = path.join(dir, entry); | |
284 | var stats = fs.statSync(filePath); | |
285 | ||
286 | if(stats.isDirectory()) { | |
287 | if(entry.substr(0, 1) == "@") { | |
288 | // When we encounter a namespace folder, augment all packages belonging to the scope | |
289 | var pkgFiles = fs.readdirSync(filePath); | |
290 | ||
291 | pkgFiles.forEach(function(entry) { | |
292 | if(stats.isDirectory()) { | |
293 | var pkgFilePath = path.join(filePath, entry); | |
294 | augmentPackageJSON(pkgFilePath, dependencies); | |
295 | } | |
296 | }); | |
297 | } else { | |
298 | augmentPackageJSON(filePath, dependencies); | |
299 | } | |
300 | } | |
301 | }); | |
302 | } | |
303 | } | |
304 | ||
305 | processDependencies("node_modules", lockObj.dependencies); | |
306 | ||
307 | fs.writeFileSync("package-lock.json", JSON.stringify(lockObj, null, 2)); | |
308 | ''; | |
309 | }; | |
310 | ||
311 | # Builds and composes an NPM package including all its dependencies | |
312 | buildNodePackage = | |
313 | { name | |
314 | , packageName | |
315 | , version | |
316 | , dependencies ? [] | |
317 | , buildInputs ? [] | |
318 | , production ? true | |
319 | , npmFlags ? "" | |
320 | , dontNpmInstall ? false | |
321 | , bypassCache ? false | |
322 | , preRebuild ? "" | |
323 | , dontStrip ? true | |
324 | , unpackPhase ? "true" | |
325 | , buildPhase ? "true" | |
326 | , ... }@args: | |
327 | ||
328 | let | |
329 | forceOfflineFlag = if bypassCache then "--offline" else "--registry http://www.example.com"; | |
330 | extraArgs = removeAttrs args [ "name" "dependencies" "buildInputs" "dontStrip" "dontNpmInstall" "preRebuild" "unpackPhase" "buildPhase" ]; | |
331 | in | |
332 | stdenv.mkDerivation ({ | |
333 | name = "node-${name}-${version}"; | |
334 | buildInputs = [ tarWrapper python nodejs ] | |
335 | ++ stdenv.lib.optional (stdenv.isLinux) utillinux | |
336 | ++ stdenv.lib.optional (stdenv.isDarwin) libtool | |
337 | ++ buildInputs; | |
338 | ||
339 | inherit dontStrip; # Stripping may fail a build for some package deployments | |
340 | inherit dontNpmInstall preRebuild unpackPhase buildPhase; | |
341 | ||
342 | compositionScript = composePackage args; | |
343 | pinpointDependenciesScript = pinpointDependenciesOfPackage args; | |
344 | ||
345 | passAsFile = [ "compositionScript" "pinpointDependenciesScript" ]; | |
346 | ||
347 | installPhase = '' | |
348 | # Create and enter a root node_modules/ folder | |
349 | mkdir -p $out/lib/node_modules | |
350 | cd $out/lib/node_modules | |
351 | ||
352 | # Compose the package and all its dependencies | |
353 | source $compositionScriptPath | |
354 | ||
355 | # Pinpoint the versions of all dependencies to the ones that are actually being used | |
356 | echo "pinpointing versions of dependencies..." | |
357 | source $pinpointDependenciesScriptPath | |
358 | ||
359 | # Patch the shebangs of the bundled modules to prevent them from | |
360 | # calling executables outside the Nix store as much as possible | |
361 | patchShebangs . | |
362 | ||
363 | # Deploy the Node.js package by running npm install. Since the | |
364 | # dependencies have been provided already by ourselves, it should not | |
365 | # attempt to install them again, which is good, because we want to make | |
366 | # it Nix's responsibility. If it needs to install any dependencies | |
367 | # anyway (e.g. because the dependency parameters are | |
368 | # incomplete/incorrect), it fails. | |
369 | # | |
370 | # The other responsibilities of NPM are kept -- version checks, build | |
371 | # steps, postprocessing etc. | |
372 | ||
373 | export HOME=$TMPDIR | |
374 | cd "${packageName}" | |
375 | runHook preRebuild | |
376 | ||
377 | ${stdenv.lib.optionalString bypassCache '' | |
378 | if [ ! -f package-lock.json ] | |
379 | then | |
380 | echo "No package-lock.json file found, reconstructing..." | |
381 | node ${reconstructPackageLock} | |
382 | fi | |
383 | ||
384 | node ${addIntegrityFieldsScript} | |
385 | ''} | |
386 | ||
387 | npm ${forceOfflineFlag} --nodedir=${nodeSources} ${npmFlags} ${stdenv.lib.optionalString production "--production"} rebuild | |
388 | ||
389 | if [ "$dontNpmInstall" != "1" ] | |
390 | then | |
391 | # NPM tries to download packages even when they already exist if npm-shrinkwrap is used. | |
392 | rm -f npm-shrinkwrap.json | |
393 | ||
394 | npm ${forceOfflineFlag} --nodedir=${nodeSources} ${npmFlags} ${stdenv.lib.optionalString production "--production"} install | |
395 | fi | |
396 | ||
397 | # Create symlink to the deployed executable folder, if applicable | |
398 | if [ -d "$out/lib/node_modules/.bin" ] | |
399 | then | |
400 | ln -s $out/lib/node_modules/.bin $out/bin | |
401 | fi | |
402 | ||
403 | # Create symlinks to the deployed manual page folders, if applicable | |
404 | if [ -d "$out/lib/node_modules/${packageName}/man" ] | |
405 | then | |
406 | mkdir -p $out/share | |
407 | for dir in "$out/lib/node_modules/${packageName}/man/"* | |
408 | do | |
409 | mkdir -p $out/share/man/$(basename "$dir") | |
410 | for page in "$dir"/* | |
411 | do | |
412 | ln -s $page $out/share/man/$(basename "$dir") | |
413 | done | |
414 | done | |
415 | fi | |
416 | ||
417 | # Run post install hook, if provided | |
418 | runHook postInstall | |
419 | ''; | |
420 | } // extraArgs); | |
421 | ||
422 | # Builds a development shell | |
423 | buildNodeShell = | |
424 | { name | |
425 | , packageName | |
426 | , version | |
427 | , src | |
428 | , dependencies ? [] | |
429 | , buildInputs ? [] | |
430 | , production ? true | |
431 | , npmFlags ? "" | |
432 | , dontNpmInstall ? false | |
433 | , bypassCache ? false | |
434 | , dontStrip ? true | |
435 | , unpackPhase ? "true" | |
436 | , buildPhase ? "true" | |
437 | , ... }@args: | |
438 | ||
439 | let | |
440 | forceOfflineFlag = if bypassCache then "--offline" else "--registry http://www.example.com"; | |
441 | ||
442 | extraArgs = removeAttrs args [ "name" "dependencies" "buildInputs" ]; | |
443 | ||
444 | nodeDependencies = stdenv.mkDerivation ({ | |
445 | name = "node-dependencies-${name}-${version}"; | |
446 | ||
447 | buildInputs = [ tarWrapper python nodejs ] | |
448 | ++ stdenv.lib.optional (stdenv.isLinux) utillinux | |
449 | ++ stdenv.lib.optional (stdenv.isDarwin) libtool | |
450 | ++ buildInputs; | |
451 | ||
452 | inherit dontStrip; # Stripping may fail a build for some package deployments | |
453 | inherit dontNpmInstall unpackPhase buildPhase; | |
454 | ||
455 | includeScript = includeDependencies { inherit dependencies; }; | |
456 | pinpointDependenciesScript = pinpointDependenciesOfPackage args; | |
457 | ||
458 | passAsFile = [ "includeScript" "pinpointDependenciesScript" ]; | |
459 | ||
460 | installPhase = '' | |
461 | mkdir -p $out/${packageName} | |
462 | cd $out/${packageName} | |
463 | ||
464 | source $includeScriptPath | |
465 | ||
466 | # Create fake package.json to make the npm commands work properly | |
467 | cp ${src}/package.json . | |
468 | chmod 644 package.json | |
469 | ${stdenv.lib.optionalString bypassCache '' | |
470 | if [ -f ${src}/package-lock.json ] | |
471 | then | |
472 | cp ${src}/package-lock.json . | |
473 | fi | |
474 | ''} | |
475 | ||
476 | # Pinpoint the versions of all dependencies to the ones that are actually being used | |
477 | echo "pinpointing versions of dependencies..." | |
478 | cd .. | |
479 | ${stdenv.lib.optionalString (builtins.substring 0 1 packageName == "@") "cd .."} | |
480 | ||
481 | source $pinpointDependenciesScriptPath | |
482 | cd ${packageName} | |
483 | ||
484 | # Patch the shebangs of the bundled modules to prevent them from | |
485 | # calling executables outside the Nix store as much as possible | |
486 | patchShebangs . | |
487 | ||
488 | export HOME=$PWD | |
489 | ||
490 | ${stdenv.lib.optionalString bypassCache '' | |
491 | if [ ! -f package-lock.json ] | |
492 | then | |
493 | echo "No package-lock.json file found, reconstructing..." | |
494 | node ${reconstructPackageLock} | |
495 | fi | |
496 | ||
497 | node ${addIntegrityFieldsScript} | |
498 | ''} | |
499 | ||
500 | npm ${forceOfflineFlag} --nodedir=${nodeSources} ${npmFlags} ${stdenv.lib.optionalString production "--production"} rebuild | |
501 | ||
502 | ${stdenv.lib.optionalString (!dontNpmInstall) '' | |
503 | # NPM tries to download packages even when they already exist if npm-shrinkwrap is used. | |
504 | rm -f npm-shrinkwrap.json | |
505 | ||
506 | npm ${forceOfflineFlag} --nodedir=${nodeSources} ${npmFlags} ${stdenv.lib.optionalString production "--production"} install | |
507 | ''} | |
508 | ||
509 | cd .. | |
510 | ${stdenv.lib.optionalString (builtins.substring 0 1 packageName == "@") "cd .."} | |
511 | ||
512 | mv ${packageName} lib | |
513 | ln -s $out/lib/node_modules/.bin $out/bin | |
514 | ''; | |
515 | } // extraArgs); | |
516 | in | |
517 | stdenv.mkDerivation { | |
518 | name = "node-shell-${name}-${version}"; | |
519 | ||
520 | buildInputs = [ python nodejs ] ++ stdenv.lib.optional (stdenv.isLinux) utillinux ++ buildInputs; | |
521 | buildCommand = '' | |
522 | mkdir -p $out/bin | |
523 | cat > $out/bin/shell <<EOF | |
524 | #! ${stdenv.shell} -e | |
525 | $shellHook | |
526 | exec ${stdenv.shell} | |
527 | EOF | |
528 | chmod +x $out/bin/shell | |
529 | ''; | |
530 | ||
531 | # Provide the dependencies in a development shell through the NODE_PATH environment variable | |
532 | inherit nodeDependencies; | |
533 | shellHook = stdenv.lib.optionalString (dependencies != []) '' | |
534 | export NODE_PATH=$nodeDependencies/lib/node_modules | |
535 | ''; | |
536 | }; | |
537 | in | |
538 | { | |
539 | buildNodeSourceDist = stdenv.lib.makeOverridable buildNodeSourceDist; | |
540 | buildNodePackage = stdenv.lib.makeOverridable buildNodePackage; | |
541 | buildNodeShell = stdenv.lib.makeOverridable buildNodeShell; | |
542 | } |