mirror of
https://github.com/jupyterhub/jupyterhub.git
synced 2025-10-07 10:04:07 +00:00
Compare commits
757 Commits
Author | SHA1 | Date | |
---|---|---|---|
![]() |
fc6cd33ce0 | ||
![]() |
b0b8e2d058 | ||
![]() |
6bfa402bfa | ||
![]() |
b51a0bba92 | ||
![]() |
2d3f962a1d | ||
![]() |
625242136a | ||
![]() |
f92560fed0 | ||
![]() |
8249ef69f0 | ||
![]() |
c63605425f | ||
![]() |
5b57900c0b | ||
![]() |
d0afdabd4c | ||
![]() |
618746fa00 | ||
![]() |
e7bc6c2ba9 | ||
![]() |
e9f86cd602 | ||
![]() |
6e8517f795 | ||
![]() |
5fa540bea1 | ||
![]() |
99f597887c | ||
![]() |
352526c36a | ||
![]() |
cbbed04eed | ||
![]() |
b2756fb18c | ||
![]() |
37b88029e4 | ||
![]() |
4b7413184e | ||
![]() |
41ef0da180 | ||
![]() |
a4a8b3fa2c | ||
![]() |
02e5984f34 | ||
![]() |
b91c5a489c | ||
![]() |
c47c3b2f9e | ||
![]() |
eaa1353dcd | ||
![]() |
b9a3b0a66a | ||
![]() |
929b805fae | ||
![]() |
082f6516a1 | ||
![]() |
1aa21f1d6c | ||
![]() |
cec9702796 | ||
![]() |
f8cbda9c3c | ||
![]() |
71aee05bc0 | ||
![]() |
772de55a0d | ||
![]() |
e6f92238b1 | ||
![]() |
db76b52e35 | ||
![]() |
e6e994e843 | ||
![]() |
284e379341 | ||
![]() |
3ce1cc63af | ||
![]() |
9945a7f7be | ||
![]() |
004c964cc1 | ||
![]() |
0f0d6d12d3 | ||
![]() |
c97e4d4e2f | ||
![]() |
53d496aff5 | ||
![]() |
032ae29066 | ||
![]() |
21caa57e7b | ||
![]() |
37ee104afa | ||
![]() |
dac75ff996 | ||
![]() |
67e06e5a18 | ||
![]() |
4cbc0bad34 | ||
![]() |
9f8c1decc4 | ||
![]() |
1244533387 | ||
![]() |
8c30724f17 | ||
![]() |
50868f5bb5 | ||
![]() |
e15b6ad52e | ||
![]() |
b194135a0f | ||
![]() |
5b8a7fd191 | ||
![]() |
be272ffb2a | ||
![]() |
8ee60ce0c7 | ||
![]() |
e553bcb7e2 | ||
![]() |
c0288ec6f6 | ||
![]() |
65b83f5f00 | ||
![]() |
dcd520179c | ||
![]() |
c830d964d5 | ||
![]() |
9e5993f1da | ||
![]() |
7ed3e0506b | ||
![]() |
7045e1116c | ||
![]() |
fb56fd406f | ||
![]() |
5489395272 | ||
![]() |
6ecda96dd6 | ||
![]() |
80ad455fc7 | ||
![]() |
21eaf0dd9f | ||
![]() |
84d2524025 | ||
![]() |
959dfb145a | ||
![]() |
998c18df42 | ||
![]() |
88b10aa2f5 | ||
![]() |
d8f5758e08 | ||
![]() |
47e45a4d3f | ||
![]() |
3e31ff4ac7 | ||
![]() |
ff30396a8e | ||
![]() |
196a7fbc65 | ||
![]() |
c66e8bb4c9 | ||
![]() |
5595146fe2 | ||
![]() |
76b688e574 | ||
![]() |
f00d0be4d6 | ||
![]() |
f9d815676f | ||
![]() |
94612d09a6 | ||
![]() |
76ed65ed82 | ||
![]() |
560bab395b | ||
![]() |
c68b846eef | ||
![]() |
5896b2c9f7 | ||
![]() |
0317fd63fa | ||
![]() |
7f6886c60f | ||
![]() |
10bdca8901 | ||
![]() |
66cb2c0f3e | ||
![]() |
0152e29946 | ||
![]() |
c6f0c07931 | ||
![]() |
51ceab9f6f | ||
![]() |
46ead8cd9d | ||
![]() |
bfb3d50936 | ||
![]() |
962307475e | ||
![]() |
80f4edcd20 | ||
![]() |
1ad4035943 | ||
![]() |
5ab735fea3 | ||
![]() |
e79cb0d376 | ||
![]() |
f728cf89c6 | ||
![]() |
8f719e21d2 | ||
![]() |
29de00ee3c | ||
![]() |
52291b0012 | ||
![]() |
e58c341290 | ||
![]() |
f988a4939e | ||
![]() |
60ee2bfc35 | ||
![]() |
42601c52cc | ||
![]() |
0679586b2c | ||
![]() |
be4201f7ee | ||
![]() |
11a73b5630 | ||
![]() |
f1efac41bf | ||
![]() |
aa6921dd5a | ||
![]() |
e94da17c3c | ||
![]() |
e2ee18fa86 | ||
![]() |
c5ec8ceba3 | ||
![]() |
3458c742cb | ||
![]() |
d1a85e53dc | ||
![]() |
d915cc3ff2 | ||
![]() |
b11c02c6e0 | ||
![]() |
49f3bb53f4 | ||
![]() |
9b7a94046b | ||
![]() |
62ef5ca2fe | ||
![]() |
028e0b0b77 | ||
![]() |
d2a42a69b0 | ||
![]() |
1f21f283df | ||
![]() |
7f35158575 | ||
![]() |
d0da677813 | ||
![]() |
a0a02688c5 | ||
![]() |
2372842b8a | ||
![]() |
7e205a9751 | ||
![]() |
e7fab5c304 | ||
![]() |
8b8b512d06 | ||
![]() |
714072dbd8 | ||
![]() |
6e8f39c22d | ||
![]() |
f3c3225124 | ||
![]() |
614bfe77d8 | ||
![]() |
1beea06ce5 | ||
![]() |
42adb44153 | ||
![]() |
d5a0202106 | ||
![]() |
3d524f2092 | ||
![]() |
409835303e | ||
![]() |
acc8d15fec | ||
![]() |
608cad6404 | ||
![]() |
571a428375 | ||
![]() |
1575adf272 | ||
![]() |
4bc6d869f3 | ||
![]() |
e5a6119505 | ||
![]() |
d80dab284d | ||
![]() |
9d556728bb | ||
![]() |
4369e2cbfa | ||
![]() |
ef4455bb67 | ||
![]() |
76c9111d80 | ||
![]() |
946ed844c5 | ||
![]() |
cceb652039 | ||
![]() |
6e988bf587 | ||
![]() |
dbc6998375 | ||
![]() |
1bdc9aa297 | ||
![]() |
73f1211286 | ||
![]() |
3fece09dda | ||
![]() |
7ad4b0c7cb | ||
![]() |
252015f50d | ||
![]() |
b3cc235c8a | ||
![]() |
47d7af8f48 | ||
![]() |
8528684dc4 | ||
![]() |
d4ce3aa731 | ||
![]() |
ec710f4d90 | ||
![]() |
14378f4cc2 | ||
![]() |
cc8e780653 | ||
![]() |
5bbf584cb7 | ||
![]() |
b5defabf49 | ||
![]() |
2d1f91e527 | ||
![]() |
1653ee77ed | ||
![]() |
10f09f4f70 | ||
![]() |
b7f277147b | ||
![]() |
f3be735eeb | ||
![]() |
3e855eb1be | ||
![]() |
98dc1f71db | ||
![]() |
703703a648 | ||
![]() |
8db8df6d7a | ||
![]() |
744430ba76 | ||
![]() |
45b858c5af | ||
![]() |
d4b5373c05 | ||
![]() |
aba55cc093 | ||
![]() |
5957a37933 | ||
![]() |
d20a33a0e4 | ||
![]() |
df35268bfe | ||
![]() |
c357d02b56 | ||
![]() |
4eb22821f2 | ||
![]() |
b92ea54eda | ||
![]() |
522ef3daea | ||
![]() |
77edffd695 | ||
![]() |
a8bc4f8a4a | ||
![]() |
66c3760b02 | ||
![]() |
fd28e224f2 | ||
![]() |
da3fedb5aa | ||
![]() |
e4e4d472b8 | ||
![]() |
bcbc68dd82 | ||
![]() |
c7df0587d2 | ||
![]() |
cd36733858 | ||
![]() |
6bf4f3b2aa | ||
![]() |
12d81ac07a | ||
![]() |
d60fa9a400 | ||
![]() |
81d423d6c6 | ||
![]() |
069b477ff3 | ||
![]() |
cf9046ea47 | ||
![]() |
71a25d4514 | ||
![]() |
2ff7d05b15 | ||
![]() |
bdb29df82a | ||
![]() |
0dbad9bd99 | ||
![]() |
2991d2d1f1 | ||
![]() |
a36a56b4ff | ||
![]() |
0e59ab003a | ||
![]() |
d67b71b7ae | ||
![]() |
8859bf8842 | ||
![]() |
4e29342711 | ||
![]() |
8a3790b01f | ||
![]() |
0d245fe4e4 | ||
![]() |
da34c6cb34 | ||
![]() |
9c0e5ba9c2 | ||
![]() |
289c3bc3c1 | ||
![]() |
3adfec0693 | ||
![]() |
137591f458 | ||
![]() |
debd297494 | ||
![]() |
10bb5ef3c0 | ||
![]() |
42e7d1a3fb | ||
![]() |
5fbd2838c9 | ||
![]() |
17dde3a2a9 | ||
![]() |
8d50554849 | ||
![]() |
493eb03345 | ||
![]() |
1beac49f4a | ||
![]() |
f230be5ede | ||
![]() |
6283e7ec83 | ||
![]() |
2438766418 | ||
![]() |
6f2e409fb9 | ||
![]() |
aa459aeb39 | ||
![]() |
9d6e8e6b6f | ||
![]() |
e882e7954c | ||
![]() |
c234463a67 | ||
![]() |
391320a590 | ||
![]() |
8648285375 | ||
![]() |
485c7b72c2 | ||
![]() |
e93cc83d58 | ||
![]() |
39b9f592b6 | ||
![]() |
1f515464fe | ||
![]() |
854d0cbb86 | ||
![]() |
87212a7414 | ||
![]() |
2338035df2 | ||
![]() |
ea132ff88d | ||
![]() |
78c14c05f3 | ||
![]() |
1d2b36e9b0 | ||
![]() |
a929ff84c7 | ||
![]() |
0d5bbc16cf | ||
![]() |
ee1fd5a469 | ||
![]() |
a702f36524 | ||
![]() |
59edc6d369 | ||
![]() |
907b77788d | ||
![]() |
914a3eaba5 | ||
![]() |
b1f048f2ef | ||
![]() |
53d76ad3a2 | ||
![]() |
7af70b92e9 | ||
![]() |
3425eca4ff | ||
![]() |
9e0bf9cd9f | ||
![]() |
3118918098 | ||
![]() |
6a995c822c | ||
![]() |
a09f535e8f | ||
![]() |
a60ac53c87 | ||
![]() |
d2c81bc1d0 | ||
![]() |
3908c6d041 | ||
![]() |
c50e1f9852 | ||
![]() |
6954e03bb4 | ||
![]() |
08eee9309e | ||
![]() |
6ed41b38ed | ||
![]() |
6b521e0b86 | ||
![]() |
1bdc66c75b | ||
![]() |
e30b2ca875 | ||
![]() |
1f3ed58570 | ||
![]() |
6a31b640c1 | ||
![]() |
ed97150311 | ||
![]() |
78eb77f157 | ||
![]() |
f152288d76 | ||
![]() |
492c5072b7 | ||
![]() |
534e251f97 | ||
![]() |
cfcd85a188 | ||
![]() |
fd3b5ebbad | ||
![]() |
1a2d5913eb | ||
![]() |
8f46d89ac0 | ||
![]() |
e82c06cf93 | ||
![]() |
392525571f | ||
![]() |
53927f0490 | ||
![]() |
ede71db11a | ||
![]() |
a2e2b1d512 | ||
![]() |
cff18992ad | ||
![]() |
b2c0b5024c | ||
![]() |
996483de94 | ||
![]() |
f4b7b85b02 | ||
![]() |
b4391d0f79 | ||
![]() |
f49cc1fcf0 | ||
![]() |
18205fbf4a | ||
![]() |
2f6ea71106 | ||
![]() |
7b6ac158cc | ||
![]() |
facf52f117 | ||
![]() |
f36796dd85 | ||
![]() |
0427f8090f | ||
![]() |
da86eaad97 | ||
![]() |
3b05135f11 | ||
![]() |
76afec8adb | ||
![]() |
06da90ac76 | ||
![]() |
7e3caf7f48 | ||
![]() |
e08552eb99 | ||
![]() |
5fb403af4b | ||
![]() |
84acdd5a7f | ||
![]() |
3e6abb7a5e | ||
![]() |
0315f986db | ||
![]() |
7735c7ddd4 | ||
![]() |
239a4c63a2 | ||
![]() |
f5bd5b7751 | ||
![]() |
287b0302d9 | ||
![]() |
44e23aad78 | ||
![]() |
606775f72d | ||
![]() |
9a6308f8d9 | ||
![]() |
0c4db2d99f | ||
![]() |
938970817c | ||
![]() |
d2a1b8e349 | ||
![]() |
4477506345 | ||
![]() |
0787489e1b | ||
![]() |
436757dd55 | ||
![]() |
a0b6d8ec6f | ||
![]() |
b92efcd7b0 | ||
![]() |
3e17b47ec3 | ||
![]() |
31c0788bd9 | ||
![]() |
dec3244758 | ||
![]() |
91e385efa7 | ||
![]() |
13313abb37 | ||
![]() |
79a51dfdce | ||
![]() |
a999ac8f07 | ||
![]() |
a3e3f24d2d | ||
![]() |
b2b85eb548 | ||
![]() |
95c5ebb090 | ||
![]() |
3d0da4f25a | ||
![]() |
bc7bb5076f | ||
![]() |
a80561bfc8 | ||
![]() |
22f86ad76c | ||
![]() |
0ae9cfa42f | ||
![]() |
ff8c4ca8a3 | ||
![]() |
ed4ed4de9d | ||
![]() |
d177b99f3a | ||
![]() |
65de8c4916 | ||
![]() |
178f9d4c51 | ||
![]() |
9433564c5b | ||
![]() |
5deba0c4ba | ||
![]() |
5234d4c7ae | ||
![]() |
1bea28026e | ||
![]() |
9a5c8ff058 | ||
![]() |
2b183c9773 | ||
![]() |
5dee864afd | ||
![]() |
6fdf931515 | ||
![]() |
d126baa443 | ||
![]() |
d1e2d593ff | ||
![]() |
3663d7c8fc | ||
![]() |
a30e6b539f | ||
![]() |
800b6a6bc5 | ||
![]() |
ca3982337e | ||
![]() |
159b3553a9 | ||
![]() |
6821e63b71 | ||
![]() |
c1c13930f7 | ||
![]() |
58f18bffff | ||
![]() |
b80906b8c8 | ||
![]() |
07aa077eae | ||
![]() |
3f74c30288 | ||
![]() |
141cb04b27 | ||
![]() |
8769864f24 | ||
![]() |
8ee72dd80f | ||
![]() |
455475724a | ||
![]() |
794be0de8e | ||
![]() |
1f633e188d | ||
![]() |
df0745985b | ||
![]() |
cad027f3fc | ||
![]() |
61a844b413 | ||
![]() |
319b404ef4 | ||
![]() |
19fb7eb7cc | ||
![]() |
cb3b0ce266 | ||
![]() |
82d8e9c433 | ||
![]() |
86ee4cad59 | ||
![]() |
add9666fcd | ||
![]() |
c93687eaad | ||
![]() |
d848873685 | ||
![]() |
c27576a41f | ||
![]() |
6d3ed95b84 | ||
![]() |
ff7cd082ff | ||
![]() |
3582ecc9cc | ||
![]() |
5f626268ef | ||
![]() |
6227f92b5f | ||
![]() |
020ba08635 | ||
![]() |
2ad175816a | ||
![]() |
3d46083dcc | ||
![]() |
dad1417b23 | ||
![]() |
9a3c2409d1 | ||
![]() |
0efb16793e | ||
![]() |
68ad36e945 | ||
![]() |
989ed216a7 | ||
![]() |
319113024d | ||
![]() |
399f7e7b80 | ||
![]() |
b4a6e5c2fe | ||
![]() |
1949ab892a | ||
![]() |
1ec34b256c | ||
![]() |
3c12a99415 | ||
![]() |
a8ced3a7ad | ||
![]() |
1af7deaeb3 | ||
![]() |
861a7c5c5e | ||
![]() |
1d02915f26 | ||
![]() |
90009f3c01 | ||
![]() |
dbce653b5e | ||
![]() |
b4443b1251 | ||
![]() |
155c76b299 | ||
![]() |
553be3e1d4 | ||
![]() |
e1e0a31afc | ||
![]() |
d78466507d | ||
![]() |
d9955a052d | ||
![]() |
2e40da09ea | ||
![]() |
490cf2dd82 | ||
![]() |
b0343ef8d8 | ||
![]() |
fb64b4f0a8 | ||
![]() |
5a747baeca | ||
![]() |
c4ce7faea6 | ||
![]() |
3a810c4fc0 | ||
![]() |
abb93ad799 | ||
![]() |
f31101432e | ||
![]() |
a2c98d016e | ||
![]() |
5581a2ba7e | ||
![]() |
1fe01ae173 | ||
![]() |
24706a1759 | ||
![]() |
182ac00e93 | ||
![]() |
ca81af2ae5 | ||
![]() |
92173c6053 | ||
![]() |
33e1a090d8 | ||
![]() |
e407808f47 | ||
![]() |
7b53330b20 | ||
![]() |
da02b024d6 | ||
![]() |
5502367832 | ||
![]() |
ddc61d2b62 | ||
![]() |
dc049a88eb | ||
![]() |
2b7a02697c | ||
![]() |
4e8acc71c6 | ||
![]() |
3bc0c18974 | ||
![]() |
3004f04a34 | ||
![]() |
e3f1fd0a16 | ||
![]() |
8367606012 | ||
![]() |
6956ffd2a9 | ||
![]() |
0b3ffe1a99 | ||
![]() |
e44ee6ed8a | ||
![]() |
45a4362bb3 | ||
![]() |
8e7df7ae7b | ||
![]() |
676a0da5ff | ||
![]() |
e802df9668 | ||
![]() |
c8e4d68978 | ||
![]() |
5ee2994504 | ||
![]() |
c194cb079e | ||
![]() |
1910bfacbd | ||
![]() |
e16ca97e1c | ||
![]() |
4bcfd52bc7 | ||
![]() |
29df06f0b5 | ||
![]() |
9ec4e6d1d1 | ||
![]() |
ce34c12349 | ||
![]() |
7b5a5541cb | ||
![]() |
731faf29c8 | ||
![]() |
bef561511f | ||
![]() |
f0b5446ec3 | ||
![]() |
629e829f8a | ||
![]() |
7c434adcb2 | ||
![]() |
3641abc70f | ||
![]() |
da790617e3 | ||
![]() |
35ba762c9c | ||
![]() |
42d9c31db7 | ||
![]() |
703af1dd1e | ||
![]() |
1dd09094a5 | ||
![]() |
b8c9717862 | ||
![]() |
06f89cb5ed | ||
![]() |
b5602028e5 | ||
![]() |
b1e45cde1e | ||
![]() |
ca117c251c | ||
![]() |
e815210cc7 | ||
![]() |
f37864cfd3 | ||
![]() |
d05d92c03a | ||
![]() |
948f4c44fd | ||
![]() |
5db76e6dcd | ||
![]() |
c944c0e54a | ||
![]() |
dd7fe85770 | ||
![]() |
b9c1831183 | ||
![]() |
5bbb292ef5 | ||
![]() |
e589b5d82a | ||
![]() |
465fb0a686 | ||
![]() |
9702c1756f | ||
![]() |
9990100f89 | ||
![]() |
a611298f43 | ||
![]() |
6a872b371e | ||
![]() |
1e298fb053 | ||
![]() |
51e1a15d63 | ||
![]() |
46e6d95364 | ||
![]() |
52c099193d | ||
![]() |
9d5784efb9 | ||
![]() |
2847c3a90c | ||
![]() |
d66f0635a3 | ||
![]() |
244ad7d38c | ||
![]() |
7fbf1826ea | ||
![]() |
b4a760234e | ||
![]() |
72a38a599d | ||
![]() |
8134d3bfbc | ||
![]() |
3df4afe7af | ||
![]() |
400c64b4ef | ||
![]() |
44dccb292f | ||
![]() |
0070e68702 | ||
![]() |
f3b1b5c7a6 | ||
![]() |
175c8d0585 | ||
![]() |
bc425a78bb | ||
![]() |
e0c4f9fc23 | ||
![]() |
2cac46fdb2 | ||
![]() |
66f8d6a626 | ||
![]() |
f163559f4a | ||
![]() |
a615f783a3 | ||
![]() |
3cafc7e49f | ||
![]() |
12ee42e8ae | ||
![]() |
9e5c837d3d | ||
![]() |
91be46784e | ||
![]() |
60a1c93801 | ||
![]() |
3a0a581782 | ||
![]() |
5cbf9399b2 | ||
![]() |
d942f52eeb | ||
![]() |
8c1620e6c5 | ||
![]() |
9fdab027da | ||
![]() |
bc32450005 | ||
![]() |
cc95d30dc1 | ||
![]() |
25ef67e8e0 | ||
![]() |
2ad1159f69 | ||
![]() |
561f4d0889 | ||
![]() |
cd0b3e05e2 | ||
![]() |
cdba57e96a | ||
![]() |
f13bd59f6f | ||
![]() |
89b0c421d5 | ||
![]() |
83ddee10ed | ||
![]() |
8a03b73086 | ||
![]() |
333b62f1fc | ||
![]() |
231d14e95d | ||
![]() |
9817610dc3 | ||
![]() |
aaf365c907 | ||
![]() |
0f93571ca5 | ||
![]() |
5b13f96162 | ||
![]() |
b41a383eae | ||
![]() |
1701149fd7 | ||
![]() |
5f8664723e | ||
![]() |
18ce8eb5a6 | ||
![]() |
d51d39728a | ||
![]() |
2255de7847 | ||
![]() |
a8c0609eb9 | ||
![]() |
66f29e0f5a | ||
![]() |
ca00c0eab0 | ||
![]() |
54baa0c31a | ||
![]() |
5d3dc509bd | ||
![]() |
9cf22e4106 | ||
![]() |
898fea9fdc | ||
![]() |
f79495e6bf | ||
![]() |
f474b31c94 | ||
![]() |
fafbe86b55 | ||
![]() |
82ad2dfbc6 | ||
![]() |
ac32ae496e | ||
![]() |
949d8d0bfa | ||
![]() |
7fd3271c9b | ||
![]() |
6267b752ae | ||
![]() |
7fcd6ad450 | ||
![]() |
dcde9f6222 | ||
![]() |
2e8ddeb114 | ||
![]() |
e07aaa603a | ||
![]() |
0bcd6adde6 | ||
![]() |
444029699a | ||
![]() |
b9bdc99c1d | ||
![]() |
c896fe05fd | ||
![]() |
424803bcd7 | ||
![]() |
9024cf1614 | ||
![]() |
a239a25ae0 | ||
![]() |
36a1ad0078 | ||
![]() |
6d696758e4 | ||
![]() |
2545cd9bb3 | ||
![]() |
096b159c23 | ||
![]() |
74958d9397 | ||
![]() |
9db18439af | ||
![]() |
2b6ad596d2 | ||
![]() |
917786f2f5 | ||
![]() |
a800496f6c | ||
![]() |
a92fee8a82 | ||
![]() |
7b1c4aedcf | ||
![]() |
572e008f1d | ||
![]() |
0379727cc0 | ||
![]() |
c9d52bea43 | ||
![]() |
263c5e838e | ||
![]() |
439e4381f0 | ||
![]() |
c34bcabcb9 | ||
![]() |
2b1bfa0ba7 | ||
![]() |
aea2eefa77 | ||
![]() |
dcde4020c2 | ||
![]() |
1225ff47be | ||
![]() |
5aaa5263fa | ||
![]() |
eca4f33afc | ||
![]() |
1e578a25d3 | ||
![]() |
41b2e6e401 | ||
![]() |
ced45d101a | ||
![]() |
03693c379e | ||
![]() |
0058ed803d | ||
![]() |
7d9a93ab5f | ||
![]() |
8a61eb1738 | ||
![]() |
cbbead3780 | ||
![]() |
146aec7e0c | ||
![]() |
f7e5904c5b | ||
![]() |
077727595c | ||
![]() |
4bfc69dc80 | ||
![]() |
8d7f55ce92 | ||
![]() |
cda7f73cfa | ||
![]() |
915664ede2 | ||
![]() |
037730761c | ||
![]() |
1d1e108e09 | ||
![]() |
6e71e617ed | ||
![]() |
9e0bb5cc71 | ||
![]() |
5fa268dab1 | ||
![]() |
1a26c1fb81 | ||
![]() |
2cc0eb885a | ||
![]() |
749b9e0997 | ||
![]() |
669dbfd449 | ||
![]() |
444f0ba00c | ||
![]() |
e46e724a70 | ||
![]() |
2e67a534cf | ||
![]() |
24c0829289 | ||
![]() |
60f5ce0ff8 | ||
![]() |
0325be3e13 | ||
![]() |
b37b13a939 | ||
![]() |
37642408a4 | ||
![]() |
9d2823e84b | ||
![]() |
ae7974564c | ||
![]() |
30c69f94c8 | ||
![]() |
47cf1915ff | ||
![]() |
9f32fc1854 | ||
![]() |
8a2eba1156 | ||
![]() |
254687e841 | ||
![]() |
aa59b1fca3 | ||
![]() |
88bff9d03d | ||
![]() |
3ca0f32ad3 | ||
![]() |
6a2876a9fa | ||
![]() |
fad6900779 | ||
![]() |
d8d58b2ebd | ||
![]() |
859dc34ea6 | ||
![]() |
8a37d2daec | ||
![]() |
41db9fe116 | ||
![]() |
8dce5a87bc | ||
![]() |
266e82755a | ||
![]() |
b237ab9e7b | ||
![]() |
7c78e6c326 | ||
![]() |
f1ed6c95f0 | ||
![]() |
2f0ce2a431 | ||
![]() |
adf3779d02 | ||
![]() |
73309b5741 | ||
![]() |
2320d59bd1 | ||
![]() |
1915ecd0c2 | ||
![]() |
d050242d0f | ||
![]() |
3d6d60b64e | ||
![]() |
fc90be8424 | ||
![]() |
1555abb2bf | ||
![]() |
8c8968c2b0 | ||
![]() |
69d0a47734 | ||
![]() |
5ae1fdf621 | ||
![]() |
c24f6b0a6a | ||
![]() |
11e32588d7 | ||
![]() |
34e44f2eed | ||
![]() |
c0464b2e47 | ||
![]() |
d686ae1ae7 | ||
![]() |
0dc3593661 | ||
![]() |
dc40cfe80e | ||
![]() |
d541c17974 | ||
![]() |
09cc8569b3 | ||
![]() |
3089d441b4 | ||
![]() |
19806899f2 | ||
![]() |
553e31235e | ||
![]() |
55323ec206 | ||
![]() |
49a5f3a654 | ||
![]() |
97c27774b1 | ||
![]() |
de11909a04 | ||
![]() |
2f15d5128e | ||
![]() |
276ef26161 | ||
![]() |
d5d315df08 | ||
![]() |
f7f82b8214 | ||
![]() |
ddece49abb | ||
![]() |
02192ee2d5 | ||
![]() |
a6b7e303df | ||
![]() |
5e5a976ea6 | ||
![]() |
c20c07ec87 | ||
![]() |
bac34e394b | ||
![]() |
2ce223c811 | ||
![]() |
e107c84162 | ||
![]() |
1cea503292 | ||
![]() |
e9bc25cce0 | ||
![]() |
8f7e25f9a1 | ||
![]() |
399def182b | ||
![]() |
f830b2a417 | ||
![]() |
cab1bca6fb | ||
![]() |
5eb7a14a33 | ||
![]() |
19da170435 | ||
![]() |
30cfdcaa83 | ||
![]() |
e9c78422b5 | ||
![]() |
844817297e | ||
![]() |
b624116be7 | ||
![]() |
38cf95523f | ||
![]() |
d6d8590acb | ||
![]() |
da460064ae | ||
![]() |
8a6de3006c | ||
![]() |
9e35ba5bef | ||
![]() |
c83777ccdc | ||
![]() |
aaad55e076 | ||
![]() |
c1e359bd38 | ||
![]() |
53f5dbd902 | ||
![]() |
9e7b0c0bfd | ||
![]() |
0aca778a9e | ||
![]() |
83af28c137 | ||
![]() |
bfbf2c0521 | ||
![]() |
09edf38a35 | ||
![]() |
e4d4e059bd | ||
![]() |
2967383654 | ||
![]() |
85f5ae1a37 | ||
![]() |
ecafe4add9 | ||
![]() |
9462511aa5 | ||
![]() |
31736eea9a | ||
![]() |
f97ef7eaac | ||
![]() |
2065099338 | ||
![]() |
d4df579fa6 | ||
![]() |
4378603e83 | ||
![]() |
40db4edc6d | ||
![]() |
ccf13979e9 | ||
![]() |
76f134c393 | ||
![]() |
77d4c1f23d | ||
![]() |
5856f46e1d | ||
![]() |
edfd1eb6cf | ||
![]() |
1ae6678360 | ||
![]() |
7794eea3fb | ||
![]() |
f51e6a1ca0 | ||
![]() |
ab00a19be1 | ||
![]() |
7742bfdda5 | ||
![]() |
f3878d8216 | ||
![]() |
d17cb637fe | ||
![]() |
5b63efe63c | ||
![]() |
54816b0a7c | ||
![]() |
41fc73db42 | ||
![]() |
984d6be542 | ||
![]() |
74a457f6b5 | ||
![]() |
137a044f96 |
@@ -19,3 +19,15 @@ jobs:
|
||||
name: smoke test jupyterhub
|
||||
command: |
|
||||
docker run --rm -it jupyterhub/jupyterhub jupyterhub --help
|
||||
- run:
|
||||
name: verify static files
|
||||
command: |
|
||||
docker run --rm -it -v $PWD/dockerfiles:/io jupyterhub/jupyterhub python3 /io/test.py
|
||||
|
||||
|
||||
# Tell CircleCI to use this workflow when it builds the site
|
||||
workflows:
|
||||
version: 2
|
||||
default:
|
||||
jobs:
|
||||
- build
|
||||
|
37
.github/ISSUE_TEMPLATE/bug_report.md
vendored
37
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@@ -1,37 +0,0 @@
|
||||
---
|
||||
name: Bug report
|
||||
about: Create a report to help us improve
|
||||
|
||||
---
|
||||
|
||||
Hi! Thanks for using JupyterHub.
|
||||
|
||||
If you are reporting an issue with JupyterHub, please use the [GitHub issue](https://github.com/jupyterhub/jupyterhub/issues) search feature to check if your issue has been asked already. If it has, please add your comments to the existing issue.
|
||||
|
||||
**Describe the bug**
|
||||
A clear and concise description of what the bug is.
|
||||
|
||||
**To Reproduce**
|
||||
Steps to reproduce the behavior:
|
||||
1. Go to '...'
|
||||
2. Click on '....'
|
||||
3. Scroll down to '....'
|
||||
4. See error
|
||||
|
||||
**Expected behavior**
|
||||
A clear and concise description of what you expected to happen.
|
||||
|
||||
**Screenshots**
|
||||
If applicable, add screenshots to help explain your problem.
|
||||
|
||||
**Desktop (please complete the following information):**
|
||||
- OS: [e.g. iOS]
|
||||
- Browser [e.g. chrome, safari]
|
||||
- Version [e.g. 22]
|
||||
|
||||
**Additional context**
|
||||
Add any other context about the problem here.
|
||||
|
||||
- Running `jupyter troubleshoot` from the command line, if possible, and posting
|
||||
its output would also be helpful.
|
||||
- Running in `--debug` mode can also be helpful for troubleshooting.
|
@@ -1,7 +0,0 @@
|
||||
---
|
||||
name: Installation and configuration issues
|
||||
about: Installation and configuration assistance
|
||||
|
||||
---
|
||||
|
||||
If you are having issues with installation or configuration, you may ask for help on the JupyterHub gitter channel or file an issue here.
|
0
.github/PULL_REQUEST_TEMPLATE/.keep
vendored
0
.github/PULL_REQUEST_TEMPLATE/.keep
vendored
2
.gitignore
vendored
2
.gitignore
vendored
@@ -24,5 +24,7 @@ MANIFEST
|
||||
.coverage.*
|
||||
htmlcov
|
||||
.idea/
|
||||
.vscode/
|
||||
.pytest_cache
|
||||
pip-wheel-metadata
|
||||
docs/source/reference/metrics.rst
|
||||
|
@@ -1,15 +1,14 @@
|
||||
repos:
|
||||
- repo: https://github.com/asottile/reorder_python_imports
|
||||
rev: v1.3.5
|
||||
rev: v1.9.0
|
||||
hooks:
|
||||
- id: reorder-python-imports
|
||||
language_version: python3.6
|
||||
- repo: https://github.com/ambv/black
|
||||
rev: 18.9b0
|
||||
- repo: https://github.com/psf/black
|
||||
rev: 19.10b0
|
||||
hooks:
|
||||
- id: black
|
||||
- repo: https://github.com/pre-commit/pre-commit-hooks
|
||||
rev: v2.1.0
|
||||
rev: v2.4.0
|
||||
hooks:
|
||||
- id: end-of-file-fixer
|
||||
- id: check-json
|
||||
|
134
.travis.yml
134
.travis.yml
@@ -1,21 +1,18 @@
|
||||
dist: bionic
|
||||
language: python
|
||||
sudo: false
|
||||
cache:
|
||||
- pip
|
||||
python:
|
||||
- 3.6
|
||||
- 3.5
|
||||
- nightly
|
||||
env:
|
||||
global:
|
||||
- ASYNC_TEST_TIMEOUT=15
|
||||
- MYSQL_HOST=127.0.0.1
|
||||
- MYSQL_TCP_PORT=13306
|
||||
|
||||
# request additional services for the jobs to access
|
||||
services:
|
||||
- postgres
|
||||
- postgresql
|
||||
- docker
|
||||
|
||||
# installing dependencies
|
||||
# install dependencies for running pytest (but not linting)
|
||||
before_install:
|
||||
- set -e
|
||||
- nvm install 6; nvm use 6
|
||||
@@ -27,68 +24,97 @@ before_install:
|
||||
unset MYSQL_UNIX_PORT
|
||||
DB=mysql bash ci/docker-db.sh
|
||||
DB=mysql bash ci/init-db.sh
|
||||
pip install 'mysql-connector-python'
|
||||
# FIXME: mysql-connector-python 8.0.16 incorrectly decodes bytes to str
|
||||
# ref: https://bugs.mysql.com/bug.php?id=94944
|
||||
pip install 'mysql-connector-python==8.0.11'
|
||||
elif [[ $JUPYTERHUB_TEST_DB_URL == postgresql* ]]; then
|
||||
psql -c "CREATE USER $PGUSER WITH PASSWORD '$PGPASSWORD';" -U postgres
|
||||
DB=postgres bash ci/init-db.sh
|
||||
pip install psycopg2-binary
|
||||
fi
|
||||
|
||||
# install general dependencies
|
||||
install:
|
||||
- pip install --upgrade pip
|
||||
- pip install --upgrade --pre -r dev-requirements.txt .
|
||||
- |
|
||||
if [[ "$MASTER_DEPENDENCIES" == "True" ]]; then
|
||||
pip install git+https://github.com/ipython/traitlets#egg=traitlets --force
|
||||
fi
|
||||
- |
|
||||
if [[ "$TEST" == "jupyter_server" ]]; then
|
||||
pip uninstall notebook --yes
|
||||
pip install jupyter_server
|
||||
fi
|
||||
- pip freeze
|
||||
|
||||
# running tests
|
||||
# run tests
|
||||
script:
|
||||
- |
|
||||
# run tests
|
||||
if [[ -z "$TEST" ]]; then
|
||||
pytest -v --maxfail=2 --cov=jupyterhub jupyterhub/tests
|
||||
fi
|
||||
- |
|
||||
# run autoformat
|
||||
if [[ "$TEST" == "lint" ]]; then
|
||||
pre-commit run --all-files
|
||||
fi
|
||||
- |
|
||||
# build docs
|
||||
if [[ "$TEST" == "docs" ]]; then
|
||||
pushd docs
|
||||
pip install --upgrade -r requirements.txt
|
||||
pip install --upgrade alabaster_jupyterhub
|
||||
make html
|
||||
popd
|
||||
fi
|
||||
- pytest -v --maxfail=2 --cov=jupyterhub jupyterhub/tests
|
||||
|
||||
# collect test coverage information
|
||||
after_success:
|
||||
- codecov
|
||||
after_failure:
|
||||
- |
|
||||
# point to auto-lint-fix
|
||||
if [[ "$TEST" == "lint" ]]; then
|
||||
echo "You can install pre-commit hooks to automatically run formatting"
|
||||
echo "on each commit with:"
|
||||
echo " pre-commit install"
|
||||
echo "or you can run by hand on staged files with"
|
||||
echo " pre-commit run"
|
||||
echo "or after-the-fact on already committed files with"
|
||||
echo " pre-commit run --all-files"
|
||||
fi
|
||||
|
||||
matrix:
|
||||
fast_finish: true
|
||||
# list the jobs
|
||||
jobs:
|
||||
include:
|
||||
- python: 3.6
|
||||
env: TEST=lint
|
||||
- python: 3.6
|
||||
env: TEST=docs
|
||||
- python: 3.6
|
||||
- name: autoformatting check
|
||||
python: 3.6
|
||||
# NOTE: It does not suffice to override to: null, [], or [""]. Travis will
|
||||
# fall back to the default if we do.
|
||||
before_install: echo "Do nothing before install."
|
||||
script:
|
||||
- pre-commit run --all-files
|
||||
after_success: echo "Do nothing after success."
|
||||
after_failure:
|
||||
- |
|
||||
echo "You can install pre-commit hooks to automatically run formatting"
|
||||
echo "on each commit with:"
|
||||
echo " pre-commit install"
|
||||
echo "or you can run by hand on staged files with"
|
||||
echo " pre-commit run"
|
||||
echo "or after-the-fact on already committed files with"
|
||||
echo " pre-commit run --all-files"
|
||||
# When we run pytest, we want to run it with python>=3.5 as well as with
|
||||
# various configurations. We increment the python version at the same time
|
||||
# as we test new configurations in order to reduce the number of test jobs.
|
||||
- name: python:3.5 + dist:xenial
|
||||
python: 3.5
|
||||
dist: xenial
|
||||
- name: python:3.6 + subdomain
|
||||
python: 3.6
|
||||
env: JUPYTERHUB_TEST_SUBDOMAIN_HOST=http://localhost.jovyan.org:8000
|
||||
- python: 3.6
|
||||
- name: python:3.7 + mysql
|
||||
python: 3.7
|
||||
env:
|
||||
- JUPYTERHUB_TEST_DB_URL=mysql+mysqlconnector://root@127.0.0.1:$MYSQL_TCP_PORT/jupyterhub
|
||||
- python: 3.6
|
||||
- name: python:3.8 + postgresql
|
||||
python: 3.8
|
||||
env:
|
||||
- JUPYTERHUB_TEST_DB_URL=postgresql://postgres@127.0.0.1/jupyterhub
|
||||
- python: 3.7
|
||||
dist: xenial
|
||||
- PGUSER=jupyterhub
|
||||
- PGPASSWORD=hub[test/:?
|
||||
# The password in url below is url-encoded with: urllib.parse.quote($PGPASSWORD, safe='')
|
||||
- JUPYTERHUB_TEST_DB_URL=postgresql://jupyterhub:hub%5Btest%2F%3A%3F@127.0.0.1/jupyterhub
|
||||
- name: python:3.8 + master dependencies
|
||||
python: 3.8
|
||||
env:
|
||||
- PGUSER=jupyterhub
|
||||
- PGPASSWORD=hub[test/:?
|
||||
# The password in url below is url-encoded with: urllib.parse.quote($PGPASSWORD, safe='')
|
||||
- JUPYTERHUB_TEST_DB_URL=postgresql://jupyterhub:hub%5Btest%2F%3A%3F@127.0.0.1/jupyterhub
|
||||
- MASTER_DEPENDENCIES=True
|
||||
- name: python:3.8 + jupyter_server
|
||||
python: 3.8
|
||||
env:
|
||||
- TEST=jupyter_server
|
||||
- JUPYTERHUB_SINGLEUSER_APP=jupyterhub.tests.mockserverapp.MockServerApp
|
||||
|
||||
- name: python:nightly
|
||||
python: nightly
|
||||
allow_failures:
|
||||
- python: nightly
|
||||
- name: python:nightly
|
||||
# https://github.com/jupyterhub/jupyterhub/issues/3141
|
||||
# The latest traitlets is close to release so it should not fail
|
||||
# - name: python:3.8 + master dependencies
|
||||
fast_finish: true
|
||||
|
@@ -8,6 +8,12 @@ for a friendly and welcoming collaborative environment.
|
||||
|
||||
## Setting up a development environment
|
||||
|
||||
<!--
|
||||
https://jupyterhub.readthedocs.io/en/stable/contributing/setup.html
|
||||
contains a lot of the same information. Should we merge the docs and
|
||||
just have this page link to that one?
|
||||
-->
|
||||
|
||||
JupyterHub requires Python >= 3.5 and nodejs.
|
||||
|
||||
As a Python project, a development install of JupyterHub follows standard practices for the basics (steps 1-2).
|
||||
@@ -60,7 +66,7 @@ pre-commit run
|
||||
|
||||
which should run any autoformatting on your code
|
||||
and tell you about any errors it couldn't fix automatically.
|
||||
You may also install [black integration](https://github.com/ambv/black#editor-integration)
|
||||
You may also install [black integration](https://github.com/psf/black#editor-integration)
|
||||
into your text editor to format code automatically.
|
||||
|
||||
If you have already committed files before setting up the pre-commit
|
||||
@@ -97,6 +103,35 @@ and other collections of tests for different components.
|
||||
When writing a new test, there should usually be a test of
|
||||
similar functionality already written and related tests should
|
||||
be added nearby.
|
||||
When in doubt, feel free to ask.
|
||||
|
||||
TODO: describe some details about fixtures, etc.
|
||||
The fixtures live in `jupyterhub/tests/conftest.py`. There are
|
||||
fixtures that can be used for JupyterHub components, such as:
|
||||
|
||||
- `app`: an instance of JupyterHub with mocked parts
|
||||
- `auth_state_enabled`: enables persisting auth_state (like authentication tokens)
|
||||
- `db`: a sqlite in-memory DB session
|
||||
- `io_loop`: a Tornado event loop
|
||||
- `event_loop`: a new asyncio event loop
|
||||
- `user`: creates a new temporary user
|
||||
- `admin_user`: creates a new temporary admin user
|
||||
- single user servers
|
||||
- `cleanup_after`: allows cleanup of single user servers between tests
|
||||
- mocked service
|
||||
- `MockServiceSpawner`: a spawner that mocks services for testing with a short poll interval
|
||||
- `mockservice`: mocked service with no external service url
|
||||
- `mockservice_url`: mocked service with a url to test external services
|
||||
|
||||
And fixtures to add functionality or spawning behavior:
|
||||
|
||||
- `admin_access`: grants admin access
|
||||
- `no_patience`: sets slow-spawning timeouts to zero
|
||||
- `slow_spawn`: enables the SlowSpawner (a spawner that takes a few seconds to start)
|
||||
- `never_spawn`: enables the NeverSpawner (a spawner that will never start)
|
||||
- `bad_spawn`: enables the BadSpawner (a spawner that fails immediately)
|
||||
- `slow_bad_spawn`: enables the SlowBadSpawner (a spawner that fails after a short delay)
|
||||
|
||||
To read more about fixtures check out the
|
||||
[pytest docs](https://docs.pytest.org/en/latest/fixture.html)
|
||||
for how to use the existing fixtures, and how to create new ones.
|
||||
|
||||
When in doubt, feel free to [ask](https://gitter.im/jupyterhub/jupyterhub).
|
||||
|
87
Dockerfile
87
Dockerfile
@@ -21,40 +21,81 @@
|
||||
# your jupyterhub_config.py will be added automatically
|
||||
# from your docker directory.
|
||||
|
||||
FROM ubuntu:18.04
|
||||
LABEL maintainer="Jupyter Project <jupyter@googlegroups.com>"
|
||||
ARG BASE_IMAGE=ubuntu:focal-20200729@sha256:6f2fb2f9fb5582f8b587837afd6ea8f37d8d1d9e41168c90f410a6ef15fa8ce5
|
||||
FROM $BASE_IMAGE AS builder
|
||||
|
||||
USER root
|
||||
|
||||
# install nodejs, utf8 locale, set CDN because default httpredir is unreliable
|
||||
ENV DEBIAN_FRONTEND noninteractive
|
||||
RUN apt-get -y update && \
|
||||
apt-get -y upgrade && \
|
||||
apt-get -y install wget git bzip2 && \
|
||||
apt-get purge && \
|
||||
apt-get clean && \
|
||||
rm -rf /var/lib/apt/lists/*
|
||||
ENV LANG C.UTF-8
|
||||
RUN apt-get update \
|
||||
&& apt-get install -yq --no-install-recommends \
|
||||
build-essential \
|
||||
ca-certificates \
|
||||
locales \
|
||||
python3-dev \
|
||||
python3-pip \
|
||||
python3-pycurl \
|
||||
nodejs \
|
||||
npm \
|
||||
&& apt-get clean \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# install Python + NodeJS with conda
|
||||
RUN wget -q https://repo.continuum.io/miniconda/Miniconda3-4.5.11-Linux-x86_64.sh -O /tmp/miniconda.sh && \
|
||||
echo 'e1045ee415162f944b6aebfe560b8fee */tmp/miniconda.sh' | md5sum -c - && \
|
||||
bash /tmp/miniconda.sh -f -b -p /opt/conda && \
|
||||
/opt/conda/bin/conda install --yes -c conda-forge \
|
||||
python=3.6 sqlalchemy tornado jinja2 traitlets requests pip pycurl \
|
||||
nodejs configurable-http-proxy && \
|
||||
/opt/conda/bin/pip install --upgrade pip && \
|
||||
rm /tmp/miniconda.sh
|
||||
ENV PATH=/opt/conda/bin:$PATH
|
||||
RUN python3 -m pip install --upgrade setuptools pip wheel
|
||||
|
||||
ADD . /src/jupyterhub
|
||||
# copy everything except whats in .dockerignore, its a
|
||||
# compromise between needing to rebuild and maintaining
|
||||
# what needs to be part of the build
|
||||
COPY . /src/jupyterhub/
|
||||
WORKDIR /src/jupyterhub
|
||||
|
||||
RUN pip install . && \
|
||||
rm -rf $PWD ~/.cache ~/.npm
|
||||
# Build client component packages (they will be copied into ./share and
|
||||
# packaged with the built wheel.)
|
||||
RUN python3 setup.py bdist_wheel
|
||||
RUN python3 -m pip wheel --wheel-dir wheelhouse dist/*.whl
|
||||
|
||||
|
||||
FROM $BASE_IMAGE
|
||||
|
||||
USER root
|
||||
|
||||
ENV DEBIAN_FRONTEND=noninteractive
|
||||
|
||||
RUN apt-get update \
|
||||
&& apt-get install -yq --no-install-recommends \
|
||||
ca-certificates \
|
||||
curl \
|
||||
gnupg \
|
||||
locales \
|
||||
python3-pip \
|
||||
python3-pycurl \
|
||||
nodejs \
|
||||
npm \
|
||||
&& apt-get clean \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
ENV SHELL=/bin/bash \
|
||||
LC_ALL=en_US.UTF-8 \
|
||||
LANG=en_US.UTF-8 \
|
||||
LANGUAGE=en_US.UTF-8
|
||||
|
||||
RUN locale-gen $LC_ALL
|
||||
|
||||
# always make sure pip is up to date!
|
||||
RUN python3 -m pip install --no-cache --upgrade setuptools pip
|
||||
|
||||
RUN npm install -g configurable-http-proxy@^4.2.0 \
|
||||
&& rm -rf ~/.npm
|
||||
|
||||
# install the wheels we built in the first stage
|
||||
COPY --from=builder /src/jupyterhub/wheelhouse /tmp/wheelhouse
|
||||
RUN python3 -m pip install --no-cache /tmp/wheelhouse/*
|
||||
|
||||
RUN mkdir -p /srv/jupyterhub/
|
||||
WORKDIR /srv/jupyterhub/
|
||||
|
||||
EXPOSE 8000
|
||||
|
||||
LABEL maintainer="Jupyter Project <jupyter@googlegroups.com>"
|
||||
LABEL org.jupyter.service="jupyterhub"
|
||||
|
||||
CMD ["jupyterhub"]
|
||||
|
33
README.md
33
README.md
@@ -10,14 +10,16 @@
|
||||
# [JupyterHub](https://github.com/jupyterhub/jupyterhub)
|
||||
|
||||
|
||||
[](https://pypi.python.org/pypi/jupyterhub)
|
||||
[](https://jupyterhub.readthedocs.org/en/latest/?badge=latest)
|
||||
[](https://travis-ci.org/jupyterhub/jupyterhub)
|
||||
[](https://circleci.com/gh/jupyterhub/jupyterhub)
|
||||
[](https://codecov.io/github/jupyterhub/jupyterhub?branch=master)
|
||||
[](https://github.com/jupyterhub/jupyterhub/issues)
|
||||
[](https://discourse.jupyter.org/c/jupyterhub)
|
||||
[](https://gitter.im/jupyterhub/jupyterhub)
|
||||
[](https://pypi.python.org/pypi/jupyterhub)
|
||||
[](https://www.npmjs.com/package/jupyterhub)
|
||||
[](https://jupyterhub.readthedocs.org/en/latest/)
|
||||
[](https://travis-ci.com/jupyterhub/jupyterhub)
|
||||
[](https://hub.docker.com/r/jupyterhub/jupyterhub/tags)
|
||||
[](https://circleci.com/gh/jupyterhub/jupyterhub)<!-- CircleCI Token: b5b65862eb2617b9a8d39e79340b0a6b816da8cc -->
|
||||
[](https://codecov.io/gh/jupyterhub/jupyterhub)
|
||||
[](https://github.com/jupyterhub/jupyterhub/issues)
|
||||
[](https://discourse.jupyter.org/c/jupyterhub)
|
||||
[](https://gitter.im/jupyterhub/jupyterhub)
|
||||
|
||||
With [JupyterHub](https://jupyterhub.readthedocs.io) you can create a
|
||||
**multi-user Hub** which spawns, manages, and proxies multiple instances of the
|
||||
@@ -72,6 +74,7 @@ for administration of the Hub and its users.
|
||||
The `nodejs-legacy` package installs the `node` executable and is currently
|
||||
required for npm to work on Debian/Ubuntu.
|
||||
|
||||
- If using the default PAM Authenticator, a [pluggable authentication module (PAM)](https://en.wikipedia.org/wiki/Pluggable_authentication_module).
|
||||
- TLS certificate and key for HTTPS communication
|
||||
- Domain name
|
||||
|
||||
@@ -145,12 +148,12 @@ To start the Hub on a specific url and port ``10.0.1.2:443`` with **https**:
|
||||
|
||||
### Authenticators
|
||||
|
||||
| Authenticator | Description |
|
||||
| --------------------------------------------------------------------------- | ------------------------------------------------- |
|
||||
| PAMAuthenticator | Default, built-in authenticator |
|
||||
| [OAuthenticator](https://github.com/jupyterhub/oauthenticator) | OAuth + JupyterHub Authenticator = OAuthenticator |
|
||||
| [ldapauthenticator](https://github.com/jupyterhub/ldapauthenticator) | Simple LDAP Authenticator Plugin for JupyterHub |
|
||||
| [kdcAuthenticator](https://github.com/bloomberg/jupyterhub-kdcauthenticator)| Kerberos Authenticator Plugin for JupyterHub |
|
||||
| Authenticator | Description |
|
||||
| ---------------------------------------------------------------------------- | ------------------------------------------------- |
|
||||
| PAMAuthenticator | Default, built-in authenticator |
|
||||
| [OAuthenticator](https://github.com/jupyterhub/oauthenticator) | OAuth + JupyterHub Authenticator = OAuthenticator |
|
||||
| [ldapauthenticator](https://github.com/jupyterhub/ldapauthenticator) | Simple LDAP Authenticator Plugin for JupyterHub |
|
||||
| [kerberosauthenticator](https://github.com/jupyterhub/kerberosauthenticator) | Kerberos Authenticator Plugin for JupyterHub |
|
||||
|
||||
### Spawners
|
||||
|
||||
@@ -162,6 +165,7 @@ To start the Hub on a specific url and port ``10.0.1.2:443`` with **https**:
|
||||
| [sudospawner](https://github.com/jupyterhub/sudospawner) | Spawn single-user servers without being root |
|
||||
| [systemdspawner](https://github.com/jupyterhub/systemdspawner) | Spawn single-user notebook servers using systemd |
|
||||
| [batchspawner](https://github.com/jupyterhub/batchspawner) | Designed for clusters using batch scheduling software |
|
||||
| [yarnspawner](https://github.com/jupyterhub/yarnspawner) | Spawn single-user notebook servers distributed on a Hadoop cluster |
|
||||
| [wrapspawner](https://github.com/jupyterhub/wrapspawner) | WrapSpawner and ProfilesSpawner enabling runtime configuration of spawners |
|
||||
|
||||
## Docker
|
||||
@@ -241,6 +245,7 @@ our JupyterHub [Gitter](https://gitter.im/jupyterhub/jupyterhub) channel.
|
||||
- [Documentation for JupyterHub's REST API](http://petstore.swagger.io/?url=https://raw.githubusercontent.com/jupyter/jupyterhub/master/docs/rest-api.yml#/default)
|
||||
- [Documentation for Project Jupyter](http://jupyter.readthedocs.io/en/latest/index.html) | [PDF](https://media.readthedocs.org/pdf/jupyter/latest/jupyter.pdf)
|
||||
- [Project Jupyter website](https://jupyter.org)
|
||||
- [Project Jupyter community](https://jupyter.org/community)
|
||||
|
||||
JupyterHub follows the Jupyter [Community Guides](https://jupyter.readthedocs.io/en/latest/community/content-community.html).
|
||||
|
||||
|
@@ -2,7 +2,7 @@
|
||||
# source this file to setup postgres and mysql
|
||||
# for local testing (as similar as possible to docker)
|
||||
|
||||
set -e
|
||||
set -eu
|
||||
|
||||
export MYSQL_HOST=127.0.0.1
|
||||
export MYSQL_TCP_PORT=${MYSQL_TCP_PORT:-13306}
|
||||
@@ -40,6 +40,15 @@ for i in {1..60}; do
|
||||
done
|
||||
$CHECK
|
||||
|
||||
case "$DB" in
|
||||
"mysql")
|
||||
;;
|
||||
"postgres")
|
||||
# create the user
|
||||
psql --user postgres -c "CREATE USER $PGUSER WITH PASSWORD '$PGPASSWORD';"
|
||||
;;
|
||||
*)
|
||||
esac
|
||||
|
||||
echo -e "
|
||||
Set these environment variables:
|
||||
|
@@ -1,7 +1,7 @@
|
||||
#!/usr/bin/env bash
|
||||
# initialize jupyterhub databases for testing
|
||||
|
||||
set -e
|
||||
set -eu
|
||||
|
||||
MYSQL="mysql --user root --host $MYSQL_HOST --port $MYSQL_TCP_PORT -e "
|
||||
PSQL="psql --user postgres -c "
|
||||
@@ -23,5 +23,5 @@ set -x
|
||||
|
||||
for SUFFIX in '' _upgrade_072 _upgrade_081 _upgrade_094; do
|
||||
$SQL "DROP DATABASE jupyterhub${SUFFIX};" 2>/dev/null || true
|
||||
$SQL "CREATE DATABASE jupyterhub${SUFFIX} ${EXTRA_CREATE};"
|
||||
$SQL "CREATE DATABASE jupyterhub${SUFFIX} ${EXTRA_CREATE:-};"
|
||||
done
|
||||
|
16
demo-image/Dockerfile
Normal file
16
demo-image/Dockerfile
Normal file
@@ -0,0 +1,16 @@
|
||||
# Demo JupyterHub Docker image
|
||||
#
|
||||
# This should only be used for demo or testing and not as a base image to build on.
|
||||
#
|
||||
# It includes the notebook package and it uses the DummyAuthenticator and the SimpleLocalProcessSpawner.
|
||||
ARG BASE_IMAGE=jupyterhub/jupyterhub-onbuild
|
||||
FROM ${BASE_IMAGE}
|
||||
|
||||
# Install the notebook package
|
||||
RUN python3 -m pip install notebook
|
||||
|
||||
# Create a demo user
|
||||
RUN useradd --create-home demo
|
||||
RUN chown demo .
|
||||
|
||||
USER demo
|
25
demo-image/README.md
Normal file
25
demo-image/README.md
Normal file
@@ -0,0 +1,25 @@
|
||||
## Demo Dockerfile
|
||||
|
||||
This is a demo JupyterHub Docker image to help you get a quick overview of what
|
||||
JupyterHub is and how it works.
|
||||
|
||||
It uses the SimpleLocalProcessSpawner to spawn new user servers and
|
||||
DummyAuthenticator for authentication.
|
||||
The DummyAuthenticator allows you to log in with any username & password and the
|
||||
SimpleLocalProcessSpawner allows starting servers without having to create a
|
||||
local user for each JupyterHub user.
|
||||
|
||||
### Important!
|
||||
|
||||
This should only be used for demo or testing purposes!
|
||||
It shouldn't be used as a base image to build on.
|
||||
|
||||
### Try it
|
||||
1. `cd` to the root of your jupyterhub repo.
|
||||
|
||||
2. Build the demo image with `docker build -t jupyterhub-demo demo-image`.
|
||||
|
||||
3. Run the demo image with `docker run -d -p 8000:8000 jupyterhub-demo`.
|
||||
|
||||
4. Visit http://localhost:8000 and login with any username and password
|
||||
5. Happy demo-ing :tada:!
|
7
demo-image/jupyterhub_config.py
Normal file
7
demo-image/jupyterhub_config.py
Normal file
@@ -0,0 +1,7 @@
|
||||
# Configuration file for jupyterhub-demo
|
||||
|
||||
c = get_config()
|
||||
|
||||
# Use DummyAuthenticator and SimpleSpawner
|
||||
c.JupyterHub.spawner_class = "simple"
|
||||
c.JupyterHub.authenticator_class = "dummy"
|
@@ -14,4 +14,7 @@ pytest-asyncio
|
||||
pytest-cov
|
||||
pytest>=3.3
|
||||
requests-mock
|
||||
# blacklist urllib3 releases affected by https://github.com/urllib3/urllib3/issues/1683
|
||||
# I *think* this should only affect testing, not production
|
||||
urllib3!=1.25.4,!=1.25.5
|
||||
virtualenv
|
||||
|
@@ -12,7 +12,7 @@ Dockerfile.alpine contains base image for jupyterhub. It does not work independ
|
||||
|
||||
* start configurable-http-proxy in another container
|
||||
* specify CONFIGPROXY_AUTH_TOKEN env in both containers
|
||||
* put both containers on the same network (e.g. docker create network jupyterhub; docker run ... --net jupyterhub)
|
||||
* put both containers on the same network (e.g. docker network create jupyterhub; docker run ... --net jupyterhub)
|
||||
* tell jupyterhub where CHP is (e.g. c.ConfigurableHTTPProxy.api_url = 'http://chp:8001')
|
||||
* tell jupyterhub not to start the proxy itself (c.ConfigurableHTTPProxy.should_start = False)
|
||||
* Use dummy authenticator for ease of testing. Update following in jupyterhub_config file
|
||||
|
9
dockerfiles/test.py
Normal file
9
dockerfiles/test.py
Normal file
@@ -0,0 +1,9 @@
|
||||
import os
|
||||
|
||||
from jupyterhub._data import DATA_FILES_PATH
|
||||
|
||||
print(f"DATA_FILES_PATH={DATA_FILES_PATH}")
|
||||
|
||||
for sub_path in ("templates", "static/components", "static/css/style.min.css"):
|
||||
path = os.path.join(DATA_FILES_PATH, sub_path)
|
||||
assert os.path.exists(path), path
|
@@ -48,6 +48,7 @@ help:
|
||||
@echo " doctest to run all doctests embedded in the documentation (if enabled)"
|
||||
@echo " coverage to run coverage check of the documentation (if enabled)"
|
||||
@echo " spelling to run spell check on documentation"
|
||||
@echo " metrics to generate documentation for metrics by inspecting the source code"
|
||||
|
||||
clean:
|
||||
rm -rf $(BUILDDIR)/*
|
||||
@@ -60,7 +61,12 @@ rest-api: source/_static/rest-api/index.html
|
||||
source/_static/rest-api/index.html: rest-api.yml node_modules
|
||||
npm run rest-api
|
||||
|
||||
html: rest-api
|
||||
metrics: source/reference/metrics.rst
|
||||
|
||||
source/reference/metrics.rst: generate-metrics.py
|
||||
python3 generate-metrics.py
|
||||
|
||||
html: rest-api metrics
|
||||
$(SPHINXBUILD) -b html $(ALLSPHINXOPTS) $(BUILDDIR)/html
|
||||
@echo
|
||||
@echo "Build finished. The HTML pages are in $(BUILDDIR)/html."
|
||||
|
@@ -1,25 +0,0 @@
|
||||
# ReadTheDocs uses the `environment.yaml` so make sure to update that as well
|
||||
# if you change the dependencies of JupyterHub in the various `requirements.txt`
|
||||
name: jhub_docs
|
||||
channels:
|
||||
- conda-forge
|
||||
dependencies:
|
||||
- nodejs
|
||||
- python=3.6
|
||||
- alembic
|
||||
- jinja2
|
||||
- pamela
|
||||
- requests
|
||||
- sqlalchemy>=1
|
||||
- tornado>=5.0
|
||||
- traitlets>=4.1
|
||||
- sphinx>=1.7
|
||||
- pip:
|
||||
- entrypoints
|
||||
- oauthlib>=2.0
|
||||
- recommonmark==0.5.0
|
||||
- async_generator
|
||||
- prometheus_client
|
||||
- attrs>=17.4.0
|
||||
- sphinx-copybutton
|
||||
- alabaster_jupyterhub
|
57
docs/generate-metrics.py
Normal file
57
docs/generate-metrics.py
Normal file
@@ -0,0 +1,57 @@
|
||||
import os
|
||||
from os.path import join
|
||||
|
||||
from pytablewriter import RstSimpleTableWriter
|
||||
from pytablewriter.style import Style
|
||||
|
||||
import jupyterhub.metrics
|
||||
|
||||
HERE = os.path.abspath(os.path.dirname(__file__))
|
||||
|
||||
|
||||
class Generator:
|
||||
@classmethod
|
||||
def create_writer(cls, table_name, headers, values):
|
||||
writer = RstSimpleTableWriter()
|
||||
writer.table_name = table_name
|
||||
writer.headers = headers
|
||||
writer.value_matrix = values
|
||||
writer.margin = 1
|
||||
[writer.set_style(header, Style(align="center")) for header in headers]
|
||||
return writer
|
||||
|
||||
def _parse_metrics(self):
|
||||
table_rows = []
|
||||
for name in dir(jupyterhub.metrics):
|
||||
obj = getattr(jupyterhub.metrics, name)
|
||||
if obj.__class__.__module__.startswith('prometheus_client.'):
|
||||
for metric in obj.describe():
|
||||
table_rows.append([metric.type, metric.name, metric.documentation])
|
||||
return table_rows
|
||||
|
||||
def prometheus_metrics(self):
|
||||
generated_directory = f"{HERE}/source/reference"
|
||||
if not os.path.exists(generated_directory):
|
||||
os.makedirs(generated_directory)
|
||||
|
||||
filename = f"{generated_directory}/metrics.rst"
|
||||
table_name = ""
|
||||
headers = ["Type", "Name", "Description"]
|
||||
values = self._parse_metrics()
|
||||
writer = self.create_writer(table_name, headers, values)
|
||||
|
||||
title = "List of Prometheus Metrics"
|
||||
underline = "============================"
|
||||
content = f"{title}\n{underline}\n{writer.dumps()}"
|
||||
with open(filename, 'w') as f:
|
||||
f.write(content)
|
||||
print(f"Generated {filename}.")
|
||||
|
||||
|
||||
def main():
|
||||
doc_generator = Generator()
|
||||
doc_generator.prometheus_metrics()
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
@@ -1,7 +1,12 @@
|
||||
# ReadTheDocs uses the `environment.yaml` so make sure to update that as well
|
||||
# if you change this file
|
||||
-r ../requirements.txt
|
||||
|
||||
alabaster_jupyterhub
|
||||
recommonmark==0.5.0
|
||||
# Temporary fix of #3021. Revert back to released autodoc-traits when
|
||||
# 0.1.0 released.
|
||||
https://github.com/jupyterhub/autodoc-traits/archive/75885ee24636efbfebfceed1043459715049cd84.zip
|
||||
pydata-sphinx-theme
|
||||
pytablewriter>=0.56
|
||||
recommonmark>=0.6
|
||||
sphinx-copybutton
|
||||
sphinx-jsonschema
|
||||
sphinx>=1.7
|
||||
|
@@ -1,13 +1,13 @@
|
||||
# see me at: http://petstore.swagger.io/?url=https://raw.githubusercontent.com/jupyter/jupyterhub/master/docs/rest-api.yml#/default
|
||||
# see me at: http://petstore.swagger.io/?url=https://raw.githubusercontent.com/jupyterhub/jupyterhub/master/docs/rest-api.yml#/default
|
||||
swagger: '2.0'
|
||||
info:
|
||||
title: JupyterHub
|
||||
description: The REST API for JupyterHub
|
||||
version: 0.9.0dev
|
||||
version: 1.2.0dev
|
||||
license:
|
||||
name: BSD-3-Clause
|
||||
schemes:
|
||||
- [http, https]
|
||||
[http, https]
|
||||
securityDefinitions:
|
||||
token:
|
||||
type: apiKey
|
||||
@@ -190,7 +190,7 @@ paths:
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- body:
|
||||
- name: body
|
||||
in: body
|
||||
schema:
|
||||
type: object
|
||||
@@ -202,34 +202,37 @@ paths:
|
||||
Timestamp of last-seen activity for this user.
|
||||
Only needed if this is not activity associated
|
||||
with using a given server.
|
||||
required: false
|
||||
servers:
|
||||
description: |
|
||||
Register activity for specific servers by name.
|
||||
The keys of this dict are the names of servers.
|
||||
The default server has an empty name ('').
|
||||
required: false
|
||||
type: object
|
||||
properties:
|
||||
'<server name>':
|
||||
description: |
|
||||
Activity for a single server.
|
||||
type: object
|
||||
required:
|
||||
- last_activity
|
||||
properties:
|
||||
last_activity:
|
||||
required: true
|
||||
type: string
|
||||
format: date-time
|
||||
description: |
|
||||
Timestamp of last-seen activity on this server.
|
||||
example:
|
||||
last_activity: '2019-02-06T12:54:14Z'
|
||||
servers:
|
||||
'':
|
||||
last_activity: '2019-02-06T12:54:14Z'
|
||||
gpu:
|
||||
last_activity: '2019-02-06T12:54:14Z'
|
||||
|
||||
example:
|
||||
last_activity: '2019-02-06T12:54:14Z'
|
||||
servers:
|
||||
'':
|
||||
last_activity: '2019-02-06T12:54:14Z'
|
||||
gpu:
|
||||
last_activity: '2019-02-06T12:54:14Z'
|
||||
responses:
|
||||
'401':
|
||||
$ref: '#/responses/Unauthorized'
|
||||
'404':
|
||||
description: No such user
|
||||
/users/{name}/server:
|
||||
post:
|
||||
summary: Start a user's single-user notebook server
|
||||
@@ -239,15 +242,19 @@ paths:
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- options:
|
||||
- name: options
|
||||
description: |
|
||||
Spawn options can be passed as a JSON body
|
||||
when spawning via the API instead of spawn form.
|
||||
The structure of the options
|
||||
will depend on the Spawner's configuration.
|
||||
The body itself will be available as `user_options` for the
|
||||
Spawner.
|
||||
in: body
|
||||
required: false
|
||||
type: object
|
||||
schema:
|
||||
type: object
|
||||
|
||||
responses:
|
||||
'201':
|
||||
description: The user's notebook server has started
|
||||
@@ -276,11 +283,14 @@ paths:
|
||||
required: true
|
||||
type: string
|
||||
- name: server_name
|
||||
description: name given to a named-server
|
||||
description: |
|
||||
name given to a named-server.
|
||||
|
||||
Note that depending on your JupyterHub infrastructure there are chracterter size limitation to `server_name`. Default spawner with K8s pod will not allow Jupyter Notebooks to be spawned with a name that contains more than 253 characters (keep in mind that the pod will be spawned with extra characters to identify the user and hub).
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- options:
|
||||
- name: options
|
||||
description: |
|
||||
Spawn options can be passed as a JSON body
|
||||
when spawning via the API instead of spawn form.
|
||||
@@ -288,7 +298,8 @@ paths:
|
||||
will depend on the Spawner's configuration.
|
||||
in: body
|
||||
required: false
|
||||
type: object
|
||||
schema:
|
||||
type: object
|
||||
responses:
|
||||
'201':
|
||||
description: The user's notebook named-server has started
|
||||
@@ -307,19 +318,30 @@ paths:
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- name: remove
|
||||
description: |
|
||||
Whether to fully remove the server, rather than just stop it.
|
||||
Removing a server deletes things like the state of the stopped server.
|
||||
- name: body
|
||||
in: body
|
||||
required: false
|
||||
type: boolean
|
||||
schema:
|
||||
type: object
|
||||
properties:
|
||||
remove:
|
||||
type: boolean
|
||||
description: |
|
||||
Whether to fully remove the server, rather than just stop it.
|
||||
Removing a server deletes things like the state of the stopped server.
|
||||
Default: false.
|
||||
responses:
|
||||
'204':
|
||||
description: The user's notebook named-server has stopped
|
||||
'202':
|
||||
description: The user's notebook named-server has not yet stopped as it is taking a while to stop
|
||||
/users/{name}/tokens:
|
||||
parameters:
|
||||
- name: name
|
||||
description: username
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
get:
|
||||
summary: List tokens for the user
|
||||
responses:
|
||||
@@ -329,25 +351,43 @@ paths:
|
||||
type: array
|
||||
items:
|
||||
$ref: '#/definitions/Token'
|
||||
'401':
|
||||
$ref: '#/responses/Unauthorized'
|
||||
'404':
|
||||
description: No such user
|
||||
post:
|
||||
summary: Create a new token for the user
|
||||
parameters:
|
||||
- name: expires_in
|
||||
type: number
|
||||
required: false
|
||||
- name: token_params
|
||||
in: body
|
||||
description: lifetime (in seconds) after which the requested token will expire.
|
||||
- name: note
|
||||
type: string
|
||||
required: false
|
||||
in: body
|
||||
description: A note attached to the token for future bookkeeping
|
||||
schema:
|
||||
type: object
|
||||
properties:
|
||||
expires_in:
|
||||
type: number
|
||||
description: lifetime (in seconds) after which the requested token will expire.
|
||||
note:
|
||||
type: string
|
||||
description: A note attached to the token for future bookkeeping
|
||||
responses:
|
||||
'201':
|
||||
description: The newly created token
|
||||
schema:
|
||||
$ref: '#/definitions/Token'
|
||||
'400':
|
||||
description: Body must be a JSON dict or empty
|
||||
/users/{name}/tokens/{token_id}:
|
||||
parameters:
|
||||
- name: name
|
||||
description: username
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- name: token_id
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
get:
|
||||
summary: Get the model for a token by id
|
||||
responses:
|
||||
@@ -361,12 +401,13 @@ paths:
|
||||
'204':
|
||||
description: The token has been deleted
|
||||
/user:
|
||||
summary: Return authenticated user's model
|
||||
description:
|
||||
parameters:
|
||||
responses:
|
||||
'200':
|
||||
description: The authenticated user's model is returned.
|
||||
get:
|
||||
summary: Return authenticated user's model
|
||||
responses:
|
||||
'200':
|
||||
description: The authenticated user's model is returned.
|
||||
schema:
|
||||
$ref: '#/definitions/User'
|
||||
/groups:
|
||||
get:
|
||||
summary: List groups
|
||||
@@ -539,14 +580,15 @@ paths:
|
||||
Logging in via this method is only available when the active Authenticator
|
||||
accepts passwords (e.g. not OAuth).
|
||||
parameters:
|
||||
- name: username
|
||||
- name: credentials
|
||||
in: body
|
||||
required: false
|
||||
type: string
|
||||
- name: password
|
||||
in: body
|
||||
required: false
|
||||
type: string
|
||||
schema:
|
||||
type: object
|
||||
properties:
|
||||
username:
|
||||
type: string
|
||||
password:
|
||||
type: string
|
||||
responses:
|
||||
'200':
|
||||
description: The new API token
|
||||
@@ -562,10 +604,10 @@ paths:
|
||||
get:
|
||||
summary: Identify a user or service from an API token
|
||||
parameters:
|
||||
- name: token
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- name: token
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
responses:
|
||||
'200':
|
||||
description: The user or service identified by the API token
|
||||
@@ -576,14 +618,14 @@ paths:
|
||||
summary: Identify a user from a cookie
|
||||
description: Used by single-user notebook servers to hand off cookie authentication to the Hub
|
||||
parameters:
|
||||
- name: cookie_name
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- name: cookie_value
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- name: cookie_name
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
- name: cookie_value
|
||||
in: path
|
||||
required: true
|
||||
type: string
|
||||
responses:
|
||||
'200':
|
||||
description: The user identified by the cookie
|
||||
@@ -618,6 +660,11 @@ paths:
|
||||
in: query
|
||||
required: true
|
||||
type: string
|
||||
responses:
|
||||
'200':
|
||||
description: Success
|
||||
'400':
|
||||
description: OAuth2Error
|
||||
/oauth2/token:
|
||||
post:
|
||||
summary: Request an OAuth2 token
|
||||
@@ -629,27 +676,27 @@ paths:
|
||||
parameters:
|
||||
- name: client_id
|
||||
description: The client id
|
||||
in: form
|
||||
in: formData
|
||||
required: true
|
||||
type: string
|
||||
- name: client_secret
|
||||
description: The client secret
|
||||
in: form
|
||||
in: formData
|
||||
required: true
|
||||
type: string
|
||||
- name: grant_type
|
||||
description: The grant type (always 'authorization_code')
|
||||
in: form
|
||||
in: formData
|
||||
required: true
|
||||
type: string
|
||||
- name: code
|
||||
description: The code provided by the authorization redirect
|
||||
in: form
|
||||
in: formData
|
||||
required: true
|
||||
type: string
|
||||
- name: redirect_uri
|
||||
description: The redirect url
|
||||
in: form
|
||||
in: formData
|
||||
required: true
|
||||
type: string
|
||||
responses:
|
||||
@@ -668,14 +715,28 @@ paths:
|
||||
post:
|
||||
summary: Shutdown the Hub
|
||||
parameters:
|
||||
- name: proxy
|
||||
- name: body
|
||||
in: body
|
||||
type: boolean
|
||||
description: Whether the proxy should be shutdown as well (default from Hub config)
|
||||
- name: servers
|
||||
in: body
|
||||
type: boolean
|
||||
description: Whether users' notebook servers should be shutdown as well (default from Hub config)
|
||||
schema:
|
||||
type: object
|
||||
properties:
|
||||
proxy:
|
||||
type: boolean
|
||||
description: Whether the proxy should be shutdown as well (default from Hub config)
|
||||
servers:
|
||||
type: boolean
|
||||
description: Whether users' notebook servers should be shutdown as well (default from Hub config)
|
||||
responses:
|
||||
'202':
|
||||
description: Shutdown successful
|
||||
'400':
|
||||
description: Unexpeced value for proxy or servers
|
||||
# Descriptions of common responses
|
||||
responses:
|
||||
NotFound:
|
||||
description: The specified resource was not found
|
||||
Unauthorized:
|
||||
description: Authentication/Authorization error
|
||||
definitions:
|
||||
User:
|
||||
type: object
|
||||
@@ -703,11 +764,10 @@ definitions:
|
||||
format: date-time
|
||||
description: Timestamp of last-seen activity from the user
|
||||
servers:
|
||||
type: object
|
||||
type: array
|
||||
description: The active servers for this user.
|
||||
items:
|
||||
schema:
|
||||
$ref: '#/definitions/Server'
|
||||
$ref: '#/definitions/Server'
|
||||
Server:
|
||||
type: object
|
||||
properties:
|
||||
@@ -745,6 +805,9 @@ definitions:
|
||||
state:
|
||||
type: object
|
||||
description: Arbitrary internal state from this server's spawner. Only available on the hub's users list or get-user-by-name method, and only if a hub admin. None otherwise.
|
||||
user_options:
|
||||
type: object
|
||||
description: User specified options for the user's spawned instance of a single-user server.
|
||||
Group:
|
||||
type: object
|
||||
properties:
|
||||
|
@@ -1,106 +1,4 @@
|
||||
div#helm-chart-schema h2,
|
||||
div#helm-chart-schema h3,
|
||||
div#helm-chart-schema h4,
|
||||
div#helm-chart-schema h5,
|
||||
div#helm-chart-schema h6 {
|
||||
font-family: courier new;
|
||||
}
|
||||
|
||||
h3, h3 ~ * {
|
||||
margin-left: 3% !important;
|
||||
}
|
||||
|
||||
h4, h4 ~ * {
|
||||
margin-left: 6% !important;
|
||||
}
|
||||
|
||||
h5, h5 ~ * {
|
||||
margin-left: 9% !important;
|
||||
}
|
||||
|
||||
h6, h6 ~ * {
|
||||
margin-left: 12% !important;
|
||||
}
|
||||
|
||||
h7, h7 ~ * {
|
||||
margin-left: 15% !important;
|
||||
}
|
||||
|
||||
img.logo {
|
||||
width:100%
|
||||
}
|
||||
|
||||
.right-next {
|
||||
float: right;
|
||||
max-width: 45%;
|
||||
overflow: auto;
|
||||
text-overflow: ellipsis;
|
||||
white-space: nowrap;
|
||||
}
|
||||
|
||||
.right-next::after{
|
||||
content: ' »';
|
||||
}
|
||||
|
||||
.left-prev {
|
||||
float: left;
|
||||
max-width: 45%;
|
||||
overflow: auto;
|
||||
text-overflow: ellipsis;
|
||||
white-space: nowrap;
|
||||
}
|
||||
|
||||
.left-prev::before{
|
||||
content: '« ';
|
||||
}
|
||||
|
||||
.prev-next-bottom {
|
||||
margin-top: 3em;
|
||||
}
|
||||
|
||||
.prev-next-top {
|
||||
margin-bottom: 1em;
|
||||
}
|
||||
|
||||
/* Sidebar TOC and headers */
|
||||
|
||||
div.sphinxsidebarwrapper div {
|
||||
margin-bottom: .8em;
|
||||
}
|
||||
div.sphinxsidebar h3 {
|
||||
font-size: 1.3em;
|
||||
padding-top: 0px;
|
||||
font-weight: 800;
|
||||
margin-left: 0px !important;
|
||||
}
|
||||
|
||||
div.sphinxsidebar p.caption {
|
||||
font-size: 1.2em;
|
||||
margin-bottom: 0px;
|
||||
margin-left: 0px !important;
|
||||
font-weight: 900;
|
||||
color: #767676;
|
||||
}
|
||||
|
||||
div.sphinxsidebar ul {
|
||||
font-size: .8em;
|
||||
margin-top: 0px;
|
||||
padding-left: 3%;
|
||||
margin-left: 0px !important;
|
||||
}
|
||||
|
||||
div.relations ul {
|
||||
font-size: 1em;
|
||||
margin-left: 0px !important;
|
||||
}
|
||||
|
||||
div#searchbox form {
|
||||
margin-left: 0px !important;
|
||||
}
|
||||
|
||||
/* body elements */
|
||||
.toctree-wrapper span.caption-text {
|
||||
color: #767676;
|
||||
font-style: italic;
|
||||
font-weight: 300;
|
||||
/* Added to avoid logo being too squeezed */
|
||||
.navbar-brand {
|
||||
height: 4rem !important;
|
||||
}
|
||||
|
Binary file not shown.
Before Width: | Height: | Size: 38 KiB After Width: | Height: | Size: 6.7 KiB |
@@ -1,16 +0,0 @@
|
||||
{# Custom template for navigation.html
|
||||
|
||||
alabaster theme does not provide blocks for titles to
|
||||
be overridden so this custom theme handles title and
|
||||
toctree for sidebar
|
||||
#}
|
||||
<h3>{{ _('Table of Contents') }}</h3>
|
||||
{{ toctree(includehidden=theme_sidebar_includehidden, collapse=theme_sidebar_collapse) }}
|
||||
{% if theme_extra_nav_links %}
|
||||
<hr />
|
||||
<ul>
|
||||
{% for text, uri in theme_extra_nav_links.items() %}
|
||||
<li class="toctree-l1"><a href="{{ uri }}">{{ text }}</a></li>
|
||||
{% endfor %}
|
||||
</ul>
|
||||
{% endif %}
|
@@ -1,30 +0,0 @@
|
||||
{% extends '!page.html' %}
|
||||
|
||||
{# Custom template for page.html
|
||||
|
||||
Alabaster theme does not provide blocks for prev/next at bottom of each page.
|
||||
This is _in addition_ to the prev/next in the sidebar. The "Prev/Next" text
|
||||
or symbols are handled by CSS classes in _static/custom.css
|
||||
#}
|
||||
|
||||
{% macro prev_next(prev, next, prev_title='', next_title='') %}
|
||||
{%- if prev %}
|
||||
<a class='left-prev' href="{{ prev.link|e }}" title="{{ _('previous chapter')}}">{{ prev_title or prev.title }}</a>
|
||||
{%- endif %}
|
||||
{%- if next %}
|
||||
<a class='right-next' href="{{ next.link|e }}" title="{{ _('next chapter')}}">{{ next_title or next.title }}</a>
|
||||
{%- endif %}
|
||||
<div style='clear:both;'></div>
|
||||
{% endmacro %}
|
||||
|
||||
|
||||
{% block body %}
|
||||
<div class='prev-next-top'>
|
||||
{{ prev_next(prev, next, 'Previous', 'Next') }}
|
||||
</div>
|
||||
|
||||
{{super()}}
|
||||
<div class='prev-next-bottom'>
|
||||
{{ prev_next(prev, next) }}
|
||||
</div>
|
||||
{% endblock %}
|
@@ -1,17 +0,0 @@
|
||||
{# Custom template for relations.html
|
||||
|
||||
alabaster theme does not provide previous/next page by default
|
||||
#}
|
||||
<div class="relations">
|
||||
<h3>Navigation</h3>
|
||||
<ul>
|
||||
<li><a href="{{ pathto(master_doc) }}">Documentation Home</a><ul>
|
||||
{%- if prev %}
|
||||
<li><a href="{{ prev.link|e }}" title="Previous">Previous topic</a></li>
|
||||
{%- endif %}
|
||||
{%- if next %}
|
||||
<li><a href="{{ next.link|e }}" title="Next">Next topic</a></li>
|
||||
{%- endif %}
|
||||
</ul>
|
||||
</ul>
|
||||
</div>
|
@@ -18,7 +18,7 @@ JupyterHub is painless, quick and with minimal user interruption.
|
||||
Read the Changelog
|
||||
==================
|
||||
|
||||
The `changelog <changelog.html>`_ contains information on what has
|
||||
The `changelog <../changelog.html>`_ contains information on what has
|
||||
changed with the new JupyterHub release, and any deprecation warnings.
|
||||
Read these notes to familiarize yourself with the coming changes. There
|
||||
might be new releases of authenticators & spawners you are using, so
|
||||
|
@@ -1,8 +1,8 @@
|
||||
.. _api-index:
|
||||
|
||||
##################
|
||||
The JupyterHub API
|
||||
##################
|
||||
##############
|
||||
JupyterHub API
|
||||
##############
|
||||
|
||||
:Release: |release|
|
||||
:Date: |today|
|
||||
|
File diff suppressed because one or more lines are too long
@@ -1,7 +1,6 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
#
|
||||
import os
|
||||
import shlex
|
||||
import sys
|
||||
|
||||
# Set paths
|
||||
@@ -19,10 +18,10 @@ extensions = [
|
||||
'sphinx.ext.napoleon',
|
||||
'autodoc_traits',
|
||||
'sphinx_copybutton',
|
||||
'sphinx-jsonschema',
|
||||
'recommonmark',
|
||||
]
|
||||
|
||||
templates_path = ['_templates']
|
||||
|
||||
# The master toctree document.
|
||||
master_doc = 'index'
|
||||
|
||||
@@ -37,7 +36,6 @@ from os.path import dirname
|
||||
docs = dirname(dirname(__file__))
|
||||
root = dirname(docs)
|
||||
sys.path.insert(0, root)
|
||||
sys.path.insert(0, os.path.join(docs, 'sphinxext'))
|
||||
|
||||
import jupyterhub
|
||||
|
||||
@@ -59,25 +57,74 @@ default_role = 'literal'
|
||||
import recommonmark
|
||||
from recommonmark.transform import AutoStructify
|
||||
|
||||
# -- Config -------------------------------------------------------------
|
||||
from jupyterhub.app import JupyterHub
|
||||
from docutils import nodes
|
||||
from sphinx.directives.other import SphinxDirective
|
||||
from contextlib import redirect_stdout
|
||||
from io import StringIO
|
||||
|
||||
# create a temp instance of JupyterHub just to get the output of the generate-config
|
||||
# and help --all commands.
|
||||
jupyterhub_app = JupyterHub()
|
||||
|
||||
|
||||
class ConfigDirective(SphinxDirective):
|
||||
"""Generate the configuration file output for use in the documentation."""
|
||||
|
||||
has_content = False
|
||||
required_arguments = 0
|
||||
optional_arguments = 0
|
||||
final_argument_whitespace = False
|
||||
option_spec = {}
|
||||
|
||||
def run(self):
|
||||
# The generated configuration file for this version
|
||||
generated_config = jupyterhub_app.generate_config_file()
|
||||
# post-process output
|
||||
home_dir = os.environ['HOME']
|
||||
generated_config = generated_config.replace(home_dir, '$HOME', 1)
|
||||
par = nodes.literal_block(text=generated_config)
|
||||
return [par]
|
||||
|
||||
|
||||
class HelpAllDirective(SphinxDirective):
|
||||
"""Print the output of jupyterhub help --all for use in the documentation."""
|
||||
|
||||
has_content = False
|
||||
required_arguments = 0
|
||||
optional_arguments = 0
|
||||
final_argument_whitespace = False
|
||||
option_spec = {}
|
||||
|
||||
def run(self):
|
||||
# The output of the help command for this version
|
||||
buffer = StringIO()
|
||||
with redirect_stdout(buffer):
|
||||
jupyterhub_app.print_help('--help-all')
|
||||
all_help = buffer.getvalue()
|
||||
# post-process output
|
||||
home_dir = os.environ['HOME']
|
||||
all_help = all_help.replace(home_dir, '$HOME', 1)
|
||||
par = nodes.literal_block(text=all_help)
|
||||
return [par]
|
||||
|
||||
|
||||
def setup(app):
|
||||
app.add_config_value('recommonmark_config', {'enable_eval_rst': True}, True)
|
||||
app.add_stylesheet('custom.css')
|
||||
app.add_css_file('custom.css')
|
||||
app.add_transform(AutoStructify)
|
||||
app.add_directive('jupyterhub-generate-config', ConfigDirective)
|
||||
app.add_directive('jupyterhub-help-all', HelpAllDirective)
|
||||
|
||||
|
||||
source_parsers = {'.md': 'recommonmark.parser.CommonMarkParser'}
|
||||
|
||||
source_suffix = ['.rst', '.md']
|
||||
# source_encoding = 'utf-8-sig'
|
||||
|
||||
# -- Options for HTML output ----------------------------------------------
|
||||
|
||||
# The theme to use for HTML and HTML Help pages.
|
||||
import alabaster_jupyterhub
|
||||
|
||||
html_theme = 'alabaster_jupyterhub'
|
||||
html_theme_path = [alabaster_jupyterhub.get_html_theme_path()]
|
||||
html_theme = 'pydata_sphinx_theme'
|
||||
|
||||
html_logo = '_static/images/logo/logo.png'
|
||||
html_favicon = '_static/images/logo/favicon.ico'
|
||||
@@ -85,31 +132,6 @@ html_favicon = '_static/images/logo/favicon.ico'
|
||||
# Paths that contain custom static files (such as style sheets)
|
||||
html_static_path = ['_static']
|
||||
|
||||
html_theme_options = {
|
||||
'show_related': True,
|
||||
'description': 'Documentation for JupyterHub',
|
||||
'github_user': 'jupyterhub',
|
||||
'github_repo': 'jupyterhub',
|
||||
'github_banner': False,
|
||||
'github_button': True,
|
||||
'github_type': 'star',
|
||||
'show_powered_by': False,
|
||||
'extra_nav_links': {
|
||||
'GitHub Repo': 'http://github.com/jupyterhub/jupyterhub',
|
||||
'Issue Tracker': 'http://github.com/jupyterhub/jupyterhub/issues',
|
||||
},
|
||||
}
|
||||
|
||||
html_sidebars = {
|
||||
'**': [
|
||||
'about.html',
|
||||
'searchbox.html',
|
||||
'navigation.html',
|
||||
'relations.html',
|
||||
'sourcelink.html',
|
||||
]
|
||||
}
|
||||
|
||||
htmlhelp_basename = 'JupyterHubdoc'
|
||||
|
||||
# -- Options for LaTeX output ---------------------------------------------
|
||||
@@ -192,14 +214,12 @@ intersphinx_mapping = {'https://docs.python.org/3/': None}
|
||||
# -- Read The Docs --------------------------------------------------------
|
||||
|
||||
on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
|
||||
if not on_rtd:
|
||||
html_theme = 'alabaster'
|
||||
else:
|
||||
if on_rtd:
|
||||
# readthedocs.org uses their theme by default, so no need to specify it
|
||||
# build rest-api, since RTD doesn't run make
|
||||
# build both metrics and rest-api, since RTD doesn't run make
|
||||
from subprocess import check_call as sh
|
||||
|
||||
sh(['make', 'rest-api'], cwd=docs)
|
||||
sh(['make', 'metrics', 'rest-api'], cwd=docs)
|
||||
|
||||
# -- Spell checking -------------------------------------------------------
|
||||
|
||||
|
21
docs/source/contributing/index.rst
Normal file
21
docs/source/contributing/index.rst
Normal file
@@ -0,0 +1,21 @@
|
||||
============
|
||||
Contributing
|
||||
============
|
||||
|
||||
We want you to contribute to JupyterHub in ways that are most exciting
|
||||
& useful to you. We value documentation, testing, bug reporting & code equally,
|
||||
and are glad to have your contributions in whatever form you wish :)
|
||||
|
||||
Our `Code of Conduct <https://github.com/jupyter/governance/blob/master/conduct/code_of_conduct.md>`_
|
||||
(`reporting guidelines <https://github.com/jupyter/governance/blob/master/conduct/reporting_online.md>`_)
|
||||
helps keep our community welcoming to as many people as possible.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
community
|
||||
setup
|
||||
docs
|
||||
tests
|
||||
roadmap
|
||||
security
|
@@ -83,7 +83,6 @@ these will be moved at a future review of the roadmap.
|
||||
- (prometheus?) API for resource monitoring
|
||||
- tracking activity on single-user servers instead of the proxy
|
||||
- notes and activity tracking per API token
|
||||
- UI for managing named servers
|
||||
|
||||
|
||||
### Later
|
||||
|
@@ -8,7 +8,7 @@ System requirements
|
||||
===================
|
||||
|
||||
JupyterHub can only run on MacOS or Linux operating systems. If you are
|
||||
using Windows, we recommend using `VirtualBox <https://virtualbox.org>`_
|
||||
using Windows, we recommend using `VirtualBox <https://virtualbox.org>`_
|
||||
or a similar system to run `Ubuntu Linux <https://ubuntu.com>`_ for
|
||||
development.
|
||||
|
||||
@@ -18,7 +18,7 @@ Install Python
|
||||
JupyterHub is written in the `Python <https://python.org>`_ programming language, and
|
||||
requires you have at least version 3.5 installed locally. If you haven’t
|
||||
installed Python before, the recommended way to install it is to use
|
||||
`miniconda <https://conda.io/miniconda.html>`_. Remember to get the ‘Python 3’ version,
|
||||
`miniconda <https://conda.io/miniconda.html>`_. Remember to get the ‘Python 3’ version,
|
||||
and **not** the ‘Python 2’ version!
|
||||
|
||||
Install nodejs
|
||||
@@ -45,7 +45,13 @@ When developing JupyterHub, you need to make changes to the code & see
|
||||
their effects quickly. You need to do a developer install to make that
|
||||
happen.
|
||||
|
||||
1. Clone the `JupyterHub git repository <https://github.com/jupyterhub/jupyterhub>`_
|
||||
.. note:: This guide does not attempt to dictate *how* development
|
||||
environements should be isolated since that is a personal preference and can
|
||||
be achieved in many ways, for example `tox`, `conda`, `docker`, etc. See this
|
||||
`forum thread <https://discourse.jupyter.org/t/thoughts-on-using-tox/3497>`_ for
|
||||
a more detailed discussion.
|
||||
|
||||
1. Clone the `JupyterHub git repository <https://github.com/jupyterhub/jupyterhub>`_
|
||||
to your computer.
|
||||
|
||||
.. code:: bash
|
||||
@@ -93,7 +99,14 @@ happen.
|
||||
python3 -m pip install -r dev-requirements.txt
|
||||
python3 -m pip install -r requirements.txt
|
||||
|
||||
5. Install the development version of JupyterHub. This lets you edit
|
||||
5. Setup a database.
|
||||
|
||||
The default database engine is ``sqlite`` so if you are just trying
|
||||
to get up and running quickly for local development that should be
|
||||
available via `python <https://docs.python.org/3.5/library/sqlite3.html>`__.
|
||||
See :doc:`/reference/database` for details on other supported databases.
|
||||
|
||||
6. Install the development version of JupyterHub. This lets you edit
|
||||
JupyterHub code in a text editor & restart the JupyterHub process to
|
||||
see your code changes immediately.
|
||||
|
||||
@@ -101,24 +114,23 @@ happen.
|
||||
|
||||
python3 -m pip install --editable .
|
||||
|
||||
6. You are now ready to start JupyterHub!
|
||||
7. You are now ready to start JupyterHub!
|
||||
|
||||
.. code:: bash
|
||||
|
||||
jupyterhub
|
||||
|
||||
7. You can access JupyterHub from your browser at
|
||||
8. You can access JupyterHub from your browser at
|
||||
``http://localhost:8000`` now.
|
||||
|
||||
Happy developing!
|
||||
|
||||
Using DummyAuthenticator & SimpleSpawner
|
||||
========================================
|
||||
Using DummyAuthenticator & SimpleLocalProcessSpawner
|
||||
====================================================
|
||||
|
||||
To simplify testing of JupyterHub, it’s helpful to use
|
||||
:class:`~jupyterhub.auth.DummyAuthenticator` instead of the default JupyterHub
|
||||
authenticator and `SimpleSpawner <https://github.com/jupyterhub/simplespawner>`_
|
||||
instead of the default spawner.
|
||||
authenticator and SimpleLocalProcessSpawner instead of the default spawner.
|
||||
|
||||
There is a sample configuration file that does this in
|
||||
``testing/jupyterhub_config.py``. To launch jupyterhub with this
|
||||
@@ -126,7 +138,6 @@ configuration:
|
||||
|
||||
.. code:: bash
|
||||
|
||||
pip install jupyterhub-simplespawner
|
||||
jupyterhub -f testing/jupyterhub_config.py
|
||||
|
||||
The default JupyterHub `authenticator
|
||||
@@ -137,15 +148,15 @@ require your system to have user accounts for each user you want to log in to
|
||||
JupyterHub as.
|
||||
|
||||
DummyAuthenticator allows you to log in with any username & password,
|
||||
while SimpleSpawner allows you to start servers without having to
|
||||
while SimpleLocalProcessSpawner allows you to start servers without having to
|
||||
create a unix user for each JupyterHub user. Together, these make it
|
||||
much easier to test JupyterHub.
|
||||
|
||||
Tip: If you are working on parts of JupyterHub that are common to all
|
||||
authenticators & spawners, we recommend using both DummyAuthenticator &
|
||||
SimpleSpawner. If you are working on just authenticator related parts,
|
||||
use only SimpleSpawner. Similarly, if you are working on just spawner
|
||||
related parts, use only DummyAuthenticator.
|
||||
SimpleLocalProcessSpawner. If you are working on just authenticator related
|
||||
parts, use only SimpleLocalProcessSpawner. Similarly, if you are working on
|
||||
just spawner related parts, use only DummyAuthenticator.
|
||||
|
||||
Troubleshooting
|
||||
===============
|
||||
|
@@ -23,27 +23,28 @@ Running the tests
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
pytest --async-test-timeout 15 -v jupyterhub/tests
|
||||
pytest -v jupyterhub/tests
|
||||
|
||||
This should display progress as it runs all the tests, printing
|
||||
information about any test failures as they occur.
|
||||
|
||||
If you wish to confirm test coverage the run tests with the `--cov` flag:
|
||||
|
||||
The ``--async-test-timeout`` parameter is used by `pytest-tornado
|
||||
<https://github.com/eugeniy/pytest-tornado#markers>`_ to set the
|
||||
asynchronous test timeout to 15 seconds rather than the default 5,
|
||||
since some of our tests take longer than 5s to execute.
|
||||
.. code-block:: bash
|
||||
|
||||
pytest -v --cov=jupyterhub jupyterhub/tests
|
||||
|
||||
#. You can also run tests in just a specific file:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
pytest --async-test-timeout 15 -v jupyterhub/tests/<test-file-name>
|
||||
pytest -v jupyterhub/tests/<test-file-name>
|
||||
|
||||
#. To run a specific test only, you can do:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
pytest --async-test-timeout 15 -v jupyterhub/tests/<test-file-name>::<test-name>
|
||||
pytest -v jupyterhub/tests/<test-file-name>::<test-name>
|
||||
|
||||
This runs the test with function name ``<test-name>`` defined in
|
||||
``<test-file-name>``. This is very useful when you are iteratively
|
||||
@@ -63,16 +64,5 @@ Troubleshooting Test Failures
|
||||
All the tests are failing
|
||||
-------------------------
|
||||
|
||||
Make sure you have completed all the steps in :ref:`contributing/setup` sucessfully, and
|
||||
Make sure you have completed all the steps in :ref:`contributing/setup` successfully, and
|
||||
can launch ``jupyterhub`` from the terminal.
|
||||
|
||||
Tests are timing out
|
||||
--------------------
|
||||
|
||||
The ``--async-test-timeout`` parameter to ``pytest`` is used by
|
||||
`pytest-tornado <https://github.com/eugeniy/pytest-tornado#markers>`_ to set
|
||||
the asynchronous test timeout to a higher value than the default of 5s,
|
||||
since some of our tests take longer than 5s to execute. If the tests
|
||||
are still timing out, try increasing that value even more. You can
|
||||
also set an environment variable ``ASYNC_TEST_TIMEOUT`` instead of
|
||||
passing ``--async-test-timeout`` to each invocation of pytest.
|
||||
|
50
docs/source/events/index.rst
Normal file
50
docs/source/events/index.rst
Normal file
@@ -0,0 +1,50 @@
|
||||
Eventlogging and Telemetry
|
||||
==========================
|
||||
|
||||
JupyterHub can be configured to record structured events from a running server using Jupyter's `Telemetry System`_. The types of events that JupyterHub emits are defined by `JSON schemas`_ listed below_
|
||||
|
||||
emitted as JSON data, defined and validated by the JSON schemas listed below.
|
||||
|
||||
|
||||
.. _logging: https://docs.python.org/3/library/logging.html
|
||||
.. _`Telemetry System`: https://github.com/jupyter/telemetry
|
||||
.. _`JSON schemas`: https://json-schema.org/
|
||||
|
||||
How to emit events
|
||||
------------------
|
||||
|
||||
Event logging is handled by its ``Eventlog`` object. This leverages Python's standing logging_ library to emit, filter, and collect event data.
|
||||
|
||||
|
||||
To begin recording events, you'll need to set two configurations:
|
||||
|
||||
1. ``handlers``: tells the EventLog *where* to route your events. This trait is a list of Python logging handlers that route events to
|
||||
2. ``allows_schemas``: tells the EventLog *which* events should be recorded. No events are emitted by default; all recorded events must be listed here.
|
||||
|
||||
Here's a basic example:
|
||||
|
||||
.. code-block::
|
||||
|
||||
import logging
|
||||
|
||||
c.EventLog.handlers = [
|
||||
logging.FileHandler('event.log'),
|
||||
]
|
||||
|
||||
c.EventLog.allowed_schemas = [
|
||||
'hub.jupyter.org/server-action'
|
||||
]
|
||||
|
||||
The output is a file, ``"event.log"``, with events recorded as JSON data.
|
||||
|
||||
|
||||
|
||||
.. _below:
|
||||
|
||||
Event schemas
|
||||
-------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
server-actions.rst
|
1
docs/source/events/server-actions.rst
Normal file
1
docs/source/events/server-actions.rst
Normal file
@@ -0,0 +1 @@
|
||||
.. jsonschema:: ../../../jupyterhub/event-schemas/server-actions/v1.yaml
|
@@ -66,7 +66,7 @@ easy to do with RStudio too.
|
||||
|
||||
### University of Colorado Boulder
|
||||
|
||||
- (CU Research Computing) CURC
|
||||
- (CU Research Computing) CURC
|
||||
- [JupyterHub User Guide](https://www.rc.colorado.edu/support/user-guide/jupyterhub.html)
|
||||
- Slurm job dispatched on Crestone compute cluster
|
||||
- log troubleshooting
|
||||
@@ -77,13 +77,17 @@ easy to do with RStudio too.
|
||||
- Earth Lab at CU
|
||||
- [Tutorial on Parallel R on JupyterHub](https://earthdatascience.org/tutorials/parallel-r-on-jupyterhub/)
|
||||
|
||||
### George Washington University
|
||||
|
||||
- [Jupyter Hub](http://go.gwu.edu/jupyter) with university single-sign-on. Deployed early 2017.
|
||||
|
||||
### HTCondor
|
||||
|
||||
- [HTCondor Python Bindings Tutorial from HTCondor Week 2017 includes information on their JupyterHub tutorials](https://research.cs.wisc.edu/htcondor/HTCondorWeek2017/presentations/TueBockelman_Python.pdf)
|
||||
|
||||
### University of Illinois
|
||||
|
||||
- https://datascience.business.illinois.edu
|
||||
- https://datascience.business.illinois.edu (currently down; checked 04/26/19)
|
||||
|
||||
### IllustrisTNG Simulation Project
|
||||
|
||||
@@ -110,7 +114,11 @@ easy to do with RStudio too.
|
||||
- [Data Science (DICE) group](https://dice.cs.uni-paderborn.de/)
|
||||
- [nbgraderutils](https://github.com/dice-group/nbgraderutils): Use JupyterHub + nbgrader + iJava kernel for online Java exercises. Used in lecture Statistical Natural Language Processing.
|
||||
|
||||
### University of Rochester CIRC
|
||||
### Penn State University
|
||||
|
||||
- [Press release](https://news.psu.edu/story/523093/2018/05/24/new-open-source-web-apps-available-students-and-faculty): "New open-source web apps available for students and faculty" (but Hub is currently down; checked 04/26/19)
|
||||
|
||||
### University of Rochester CIRC
|
||||
|
||||
- [JupyterHub Userguide](https://info.circ.rochester.edu/Web_Applications/JupyterHub.html) - Slurm, beehive
|
||||
|
||||
@@ -126,7 +134,7 @@ easy to do with RStudio too.
|
||||
|
||||
- Educational Technology Services - Paul Jamason
|
||||
- [jupyterhub.ucsd.edu](https://jupyterhub.ucsd.edu)
|
||||
|
||||
|
||||
### TACC University of Texas
|
||||
|
||||
### Texas A&M
|
||||
@@ -134,7 +142,10 @@ easy to do with RStudio too.
|
||||
- Kristen Thyng - Oceanography
|
||||
- [Teaching with JupyterHub and nbgrader](http://kristenthyng.com/blog/2016/09/07/jupyterhub+nbgrader/)
|
||||
|
||||
|
||||
### Elucidata
|
||||
- What's new in Jupyter Notebooks @[Elucidata](https://elucidata.io/):
|
||||
- Using Jupyter Notebooks with Jupyterhub on GCP, managed by GKE
|
||||
- https://medium.com/elucidata/why-you-should-be-using-a-jupyter-notebook-8385a4ccd93d
|
||||
|
||||
## Service Providers
|
||||
|
||||
@@ -160,6 +171,10 @@ easy to do with RStudio too.
|
||||
- https://getcarina.com/blog/learning-how-to-whale/
|
||||
- http://carolynvanslyck.com/talk/carina/jupyterhub/#/
|
||||
|
||||
### Hadoop
|
||||
|
||||
- [Deploying JupyterHub on Hadoop](https://jupyterhub-on-hadoop.readthedocs.io)
|
||||
|
||||
|
||||
## Miscellaneous
|
||||
|
||||
|
@@ -4,23 +4,23 @@ The default Authenticator uses [PAM][] to authenticate system users with
|
||||
their username and password. With the default Authenticator, any user
|
||||
with an account and password on the system will be allowed to login.
|
||||
|
||||
## Create a whitelist of users
|
||||
## Create a set of allowed users
|
||||
|
||||
You can restrict which users are allowed to login with a whitelist,
|
||||
`Authenticator.whitelist`:
|
||||
You can restrict which users are allowed to login with a set,
|
||||
`Authenticator.allowed_users`:
|
||||
|
||||
|
||||
```python
|
||||
c.Authenticator.whitelist = {'mal', 'zoe', 'inara', 'kaylee'}
|
||||
c.Authenticator.allowed_users = {'mal', 'zoe', 'inara', 'kaylee'}
|
||||
```
|
||||
|
||||
Users in the whitelist are added to the Hub database when the Hub is
|
||||
Users in the `allowed_users` set are added to the Hub database when the Hub is
|
||||
started.
|
||||
|
||||
## Configure admins (`admin_users`)
|
||||
|
||||
Admin users of JupyterHub, `admin_users`, can add and remove users from
|
||||
the user `whitelist`. `admin_users` can take actions on other users'
|
||||
the user `allowed_users` set. `admin_users` can take actions on other users'
|
||||
behalf, such as stopping and restarting their servers.
|
||||
|
||||
A set of initial admin users, `admin_users` can configured be as follows:
|
||||
@@ -28,7 +28,7 @@ A set of initial admin users, `admin_users` can configured be as follows:
|
||||
```python
|
||||
c.Authenticator.admin_users = {'mal', 'zoe'}
|
||||
```
|
||||
Users in the admin list are automatically added to the user `whitelist`,
|
||||
Users in the admin set are automatically added to the user `allowed_users` set,
|
||||
if they are not already present.
|
||||
|
||||
Each authenticator may have different ways of determining whether a user is an
|
||||
@@ -53,12 +53,12 @@ sure your users know if admin_access is enabled.**
|
||||
|
||||
Users can be added to and removed from the Hub via either the admin
|
||||
panel or the REST API. When a user is **added**, the user will be
|
||||
automatically added to the whitelist and database. Restarting the Hub
|
||||
will not require manually updating the whitelist in your config file,
|
||||
automatically added to the allowed users set and database. Restarting the Hub
|
||||
will not require manually updating the allowed users set in your config file,
|
||||
as the users will be loaded from the database.
|
||||
|
||||
After starting the Hub once, it is not sufficient to **remove** a user
|
||||
from the whitelist in your config file. You must also remove the user
|
||||
from the allowed users set in your config file. You must also remove the user
|
||||
from the Hub's database, either by deleting the user from JupyterHub's
|
||||
admin page, or you can clear the `jupyterhub.sqlite` database and start
|
||||
fresh.
|
||||
|
36
docs/source/getting-started/faq.md
Normal file
36
docs/source/getting-started/faq.md
Normal file
@@ -0,0 +1,36 @@
|
||||
# Frequently asked questions
|
||||
|
||||
|
||||
### How do I share links to notebooks?
|
||||
|
||||
In short, where you see `/user/name/notebooks/foo.ipynb` use `/hub/user-redirect/notebooks/foo.ipynb` (replace `/user/name` with `/hub/user-redirect`).
|
||||
|
||||
Sharing links to notebooks is a common activity,
|
||||
and can look different based on what you mean.
|
||||
Your first instinct might be to copy the URL you see in the browser,
|
||||
e.g. `hub.jupyter.org/user/yourname/notebooks/coolthing.ipynb`.
|
||||
However, let's break down what this URL means:
|
||||
|
||||
`hub.jupyter.org/user/yourname/` is the URL prefix handled by *your server*,
|
||||
which means that sharing this URL is asking the person you share the link with
|
||||
to come to *your server* and look at the exact same file.
|
||||
In most circumstances, this is forbidden by permissions because the person you share with does not have access to your server.
|
||||
What actually happens when someone visits this URL will depend on whether your server is running and other factors.
|
||||
|
||||
But what is our actual goal?
|
||||
A typical situation is that you have some shared or common filesystem,
|
||||
such that the same path corresponds to the same document
|
||||
(either the exact same document or another copy of it).
|
||||
Typically, what folks want when they do sharing like this
|
||||
is for each visitor to open the same file *on their own server*,
|
||||
so Breq would open `/user/breq/notebooks/foo.ipynb` and
|
||||
Seivarden would open `/user/seivarden/notebooks/foo.ipynb`, etc.
|
||||
|
||||
JupyterHub has a special URL that does exactly this!
|
||||
It's called `/hub/user-redirect/...` and after the visitor logs in,
|
||||
So if you replace `/user/yourname` in your URL bar
|
||||
with `/hub/user-redirect` any visitor should get the same
|
||||
URL on their own server, rather than visiting yours.
|
||||
|
||||
In JupyterLab 2.0, this should also be the result of the "Copy Shareable Link"
|
||||
action in the file browser.
|
@@ -1,5 +1,10 @@
|
||||
Getting Started
|
||||
===============
|
||||
Get Started
|
||||
===========
|
||||
|
||||
This section covers how to configure and customize JupyterHub for your
|
||||
needs. It contains information about authentication, networking, security, and
|
||||
other topics that are relevant to individuals or organizations deploying their
|
||||
own JupyterHub.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
@@ -10,3 +15,5 @@ Getting Started
|
||||
authenticators-users-basics
|
||||
spawners-basics
|
||||
services-basics
|
||||
faq
|
||||
institutional-faq
|
||||
|
266
docs/source/getting-started/institutional-faq.md
Normal file
266
docs/source/getting-started/institutional-faq.md
Normal file
@@ -0,0 +1,266 @@
|
||||
# Institutional FAQ
|
||||
|
||||
This page contains common questions from users of JupyterHub,
|
||||
broken down by their roles within organizations.
|
||||
|
||||
## For all
|
||||
|
||||
### Is it appropriate for adoption within a larger institutional context?
|
||||
|
||||
Yes! JupyterHub has been used at-scale for large pools of users, as well
|
||||
as complex and high-performance computing. For example, UC Berkeley uses
|
||||
JupyterHub for its Data Science Education Program courses (serving over
|
||||
3,000 students). The Pangeo project uses JupyterHub to provide access
|
||||
to scalable cloud computing with Dask. JupyterHub is stable customizable
|
||||
to the use-cases of large organizations.
|
||||
|
||||
### I keep hearing about Jupyter Notebook, JupyterLab, and now JupyterHub. What’s the difference?
|
||||
|
||||
Here is a quick breakdown of these three tools:
|
||||
|
||||
* **The Jupyter Notebook** is a document specification (the `.ipynb`) file that interweaves
|
||||
narrative text with code cells and their outputs. It is also a graphical interface
|
||||
that allows users to edit these documents. There are also several other graphical interfaces
|
||||
that allow users to edit the `.ipynb` format (nteract, Jupyter Lab, Google Colab, Kaggle, etc).
|
||||
* **JupyterLab** is a flexible and extendible user interface for interactive computing. It
|
||||
has several extensions that are tailored for using Jupyter Notebooks, as well as extensions
|
||||
for other parts of the data science stack.
|
||||
* **JupyterHub** is an application that manages interactive computing sessions for **multiple users**.
|
||||
It also connects them with infrastructure those users wish to access. It can provide
|
||||
remote access to Jupyter Notebooks and Jupyter Lab for many people.
|
||||
|
||||
## For management
|
||||
|
||||
### Briefly, what problem does JupyterHub solve for us?
|
||||
|
||||
JupyterHub provides a shared platform for data science and collaboration.
|
||||
It allows users to utilize familiar data science workflows (such as the scientific python stack,
|
||||
the R tidyverse, and Jupyter Notebooks) on institutional infrastructure. It also allows administrators
|
||||
some control over access to resources, security, environments, and authentication.
|
||||
|
||||
### Is JupyterHub mature? Why should we trust it?
|
||||
|
||||
Yes - the core JupyterHub application recently
|
||||
reached 1.0 status, and is considered stable and performant for most institutions.
|
||||
JupyterHub has also been deployed (along with other tools) to work on
|
||||
scalable infrastructure, large datasets, and high-performance computing.
|
||||
|
||||
### Who else uses JupyterHub?
|
||||
|
||||
JupyterHub is used at a variety of institutions in academia,
|
||||
industry, and government research labs. It is most-commonly used by two kinds of groups:
|
||||
|
||||
* Small teams (e.g., data science teams, research labs, or collaborative projects) to provide a
|
||||
shared resource for interactive computing, collaboration, and analytics.
|
||||
* Large teams (e.g., a department, a large class, or a large group of remote users) to provide
|
||||
access to organizational hardware, data, and analytics environments at scale.
|
||||
|
||||
Here are a sample of organizations that use JupyterHub:
|
||||
|
||||
* **Universities and colleges**: UC Berkeley, UC San Diego, Cal Poly SLO, Harvard University, University of Chicago,
|
||||
University of Oslo, University of Sheffield, Université Paris Sud, University of Versailles
|
||||
* **Research laboratories**: NASA, NCAR, NOAA, the Large Synoptic Survey Telescope, Brookhaven National Lab,
|
||||
Minnesota Supercomputing Institute, ALCF, CERN, Lawrence Livermore National Laboratory
|
||||
* **Online communities**: Pangeo, Quantopian, mybinder.org, MathHub, Open Humans
|
||||
* **Computing infrastructure providers**: NERSC, San Diego Supercomputing Center, Compute Canada
|
||||
* **Companies**: Capital One, SANDVIK code, Globus
|
||||
|
||||
See the [Gallery of JupyterHub deployments](../gallery-jhub-deployments.md) for
|
||||
a more complete list of JupyterHub deployments at institutions.
|
||||
|
||||
### How does JupyterHub compare with hosted products, like Google Colaboratory, RStudio.cloud, or Anaconda Enterprise?
|
||||
|
||||
JupyterHub puts you in control of your data, infrastructure, and coding environment.
|
||||
In addition, it is vendor neutral, which reduces lock-in to a particular vendor or service.
|
||||
JupyterHub provides access to interactive computing environments in the cloud (similar to each of these services).
|
||||
Compared with the tools above, it is more flexible, more customizable, free, and
|
||||
gives administrators more control over their setup and hardware.
|
||||
|
||||
Because JupyterHub is an open-source, community-driven tool, it can be extended and
|
||||
modified to fit an institution's needs. It plays nicely with the open source data science
|
||||
stack, and can serve a variety of computing enviroments, user interfaces, and
|
||||
computational hardware. It can also be deployed anywhere - on enterprise cloud infrastructure, on
|
||||
High-Performance-Computing machines, on local hardware, or even on a single laptop, which
|
||||
is not possible with most other tools for shared interactive computing.
|
||||
|
||||
## For IT
|
||||
|
||||
### How would I set up JupyterHub on institutional hardware?
|
||||
|
||||
That depends on what kind of hardware you've got. JupyterHub is flexible enough to be deployed
|
||||
on a variety of hardware, including in-room hardware, on-prem clusters, cloud infrastructure,
|
||||
etc.
|
||||
|
||||
The most common way to set up a JupyterHub is to use a JupyterHub distribution, these are pre-configured
|
||||
and opinionated ways to set up a JupyterHub on particular kinds of infrastructure. The two distributions
|
||||
that we currently suggest are:
|
||||
|
||||
* [Zero to JupyterHub for Kubernetes](https://z2jh.jupyter.org) is a scalable JupyterHub deployment and
|
||||
guide that runs on Kubernetes. Better for larger or dynamic user groups (50-10,000) or more complex
|
||||
compute/data needs.
|
||||
* [The Littlest JupyterHub](https://tljh.jupyter.org) is a lightweight JupyterHub that runs on a single
|
||||
single machine (in the cloud or under your desk). Better for smaller usergroups (4-80) or more
|
||||
lightweight computational resources.
|
||||
|
||||
|
||||
### Does JupyterHub run well in the cloud?
|
||||
|
||||
Yes - most deployments of JupyterHub are run via cloud infrastructure and on a variety of cloud providers.
|
||||
Depending on the distribution of JupyterHub that you'd like to use, you can also connect your JupyterHub
|
||||
deployment with a number of other cloud-native services so that users have access to other resources from
|
||||
their interactive computing sessions.
|
||||
|
||||
For example, if you use the [Zero to JupyterHub for Kubernetes](https://z2jh.jupyter.org) distribution,
|
||||
you'll be able to utilize container-based workflows of other technologies such as the [dask-kubernetes](https://kubernetes.dask.org/en/latest/)
|
||||
project for distributed computing.
|
||||
|
||||
The Z2JH Helm Chart also has some functionality built in for auto-scaling your cluster up and down
|
||||
as more resources are needed - allowing you to utilize the benefits of a flexible cloud-based deployment.
|
||||
|
||||
### Is JupyterHub secure?
|
||||
|
||||
The short answer: yes. JupyterHub as a standalone application has been battle-tested at an institutional
|
||||
level for several years, and makes a number of "default" security decisions that are reasonable for most
|
||||
users.
|
||||
|
||||
* For security considerations in the base JupyterHub application,
|
||||
[see the JupyterHub security page](https://jupyterhub.readthedocs.io/en/stable/reference/websecurity.html)
|
||||
* For security considerations when deploying JupyterHub on Kubernetes, see the
|
||||
[JupyterHub on Kubernetes security page](https://zero-to-jupyterhub.readthedocs.io/en/latest/security.html).
|
||||
|
||||
The longer answer: it depends on your deployment. Because JupyterHub is very flexible, it can be used
|
||||
in a variety of deployment setups. This often entails connecting your JupyterHub to **other** infrastructure
|
||||
(such as a [Dask Gateway service](https://gateway.dask.org/)). There are many security decisions to be made
|
||||
in these cases, and the security of your JupyterHub deployment will often depend on these decisions.
|
||||
|
||||
If you are worried about security, don't hesitate to reach out to the JupyterHub community in the
|
||||
[Jupyter Community Forum](https://discourse.jupyter.org/c/jupyterhub). This community of practice has many
|
||||
individuals with experience running secure JupyterHub deployments.
|
||||
|
||||
|
||||
### Does JupyterHub provide computing or data infrastructure?
|
||||
|
||||
No - JupyterHub manages user sessions and can *control* computing infrastructure, but it does not provide these
|
||||
things itself. You are expected to run JupyterHub on your own infrastructure (local or in the cloud). Moreover,
|
||||
JupyterHub has no internal concept of "data", but is designed to be able to communicate with data repositories
|
||||
(again, either locally or remotely) for use within interactive computing sessions.
|
||||
|
||||
|
||||
### How do I manage users?
|
||||
|
||||
JupyterHub offers a few options for managing your users. Upon setting up a JupyterHub, you can choose what
|
||||
kind of **authentication** you'd like to use. For example, you can have users sign up with an institutional
|
||||
email address, or choose a username / password when they first log-in, or offload authentication onto
|
||||
another service such as an organization's OAuth.
|
||||
|
||||
The users of a JupyterHub are stored locally, and can be modified manually by an administrator of the JupyterHub.
|
||||
Moreover, the *active* users on a JupyterHub can be found on the administrator's page. This page
|
||||
gives you the abiltiy to stop or restart kernels, inspect user filesystems, and even take over user
|
||||
sessions to assist them with debugging.
|
||||
|
||||
### How do I manage software environments?
|
||||
|
||||
A key benefit of JupyterHub is the ability for an administrator to define the environment(s) that users
|
||||
have access to. There are many ways to do this, depending on what kind of infrastructure you're using for
|
||||
your JupyterHub.
|
||||
|
||||
For example, **The Littlest JupyterHub** runs on a single VM. In this case, the administrator defines
|
||||
an environment by installing packages to a shared folder that exists on the path of all users. The
|
||||
**JupyterHub for Kubernetes** deployment uses Docker images to define environments. You can create your
|
||||
own list of Docker images that users can select from, and can also control things like the amount of
|
||||
RAM available to users, or the types of machines that their sessions will use in the cloud.
|
||||
|
||||
### How does JupyterHub manage computational resources?
|
||||
|
||||
For interactive computing sessions, JupyterHub controls computational resources via a **spawner**.
|
||||
Spawners define how a new user session is created, and are customized for particular kinds of
|
||||
infrastructure. For example, the KubeSpawner knows how to control a Kubernetes deployment
|
||||
to create new pods when users log in.
|
||||
|
||||
For more sophisticated computational resources (like distributed computing), JupyterHub can
|
||||
connect with other infrastructure tools (like Dask or Spark). This allows users to control
|
||||
scalable or high-performance resources from within their JupyterHub sessions. The logic of
|
||||
how those resources are controlled is taken care of by the non-JupyterHub application.
|
||||
|
||||
|
||||
### Can JupyterHub be used with my high-performance computing resources?
|
||||
|
||||
Yes - JupyterHub can provide access to many kinds of computing infrastructure.
|
||||
Especially when combined with other open-source schedulers such as Dask, you can manage fairly
|
||||
complex computing infrastructure from the interactive sessions of a JupyterHub. For example
|
||||
[see the Dask HPC page](https://docs.dask.org/en/latest/setup/hpc.html).
|
||||
|
||||
### How much resources do user sessions take?
|
||||
|
||||
This is highly configurable by the administrator. If you wish for your users to have simple
|
||||
data analytics environments for prototyping and light data exploring, you can restrict their
|
||||
memory and CPU based on the resources that you have available. If you'd like your JupyterHub
|
||||
to serve as a gateway to high-performance compute or data resources, you may increase the
|
||||
resources available on user machines, or connect them with computing infrastructure elsewhere.
|
||||
|
||||
### Can I customize the look and feel of a JupyterHub?
|
||||
|
||||
JupyterHub provides some customization of the graphics displayed to users. The most common
|
||||
modification is to add custom branding to the JupyterHub login page, loading pages, and
|
||||
various elements that persist across all pages (such as headers).
|
||||
|
||||
## For Technical Leads
|
||||
|
||||
### Will JupyterHub “just work” with our team's interactive computing setup?
|
||||
|
||||
Depending on the complexity of your setup, you'll have different experiences with "out of the box"
|
||||
distributions of JupyterHub. If all of the resources you need will fit on a single VM, then
|
||||
[The Littlest JupyterHub](https://tljh.jupyter.org) should get you up-and-running within
|
||||
a half day or so. For more complex setups, such as scalable Kubernetes clusters or access
|
||||
to high-performance computing and data, it will require more time and expertise with
|
||||
the technologies your JupyterHub will use (e.g., dev-ops knowledge with cloud computing).
|
||||
|
||||
In general, the base JupyterHub deployment is not the bottleneck for setup, it is connecting
|
||||
your JupyterHub with the various services and tools that you wish to provide to your users.
|
||||
|
||||
|
||||
### How well does JupyterHub scale? What are JupyterHub's limitations?
|
||||
|
||||
JupyterHub works well at both a small scale (e.g., a single VM or machine) as well as a
|
||||
high scale (e.g., a scalable Kubernetes cluster). It can be used for teams as small a 2, and
|
||||
for user bases as large as 10,000. The scalability of JupyterHub largely depends on the
|
||||
infrastructure on which it is deployed. JupyterHub has been designed to be lightweight and
|
||||
flexible, so you can tailor your JupyterHub deployment to your needs.
|
||||
|
||||
|
||||
### Is JupyterHub resilient? What happens when a machine goes down?
|
||||
|
||||
For JupyterHubs that are deployed in a containerized environment (e.g., Kubernetes), it is
|
||||
possible to configure the JupyterHub to be fairly resistant to failures in the system.
|
||||
For example, if JupyterHub fails, then user sessions will not be affected (though new
|
||||
users will not be able to log in). When a JupyterHub process is restarted, it should
|
||||
seamlessly connect with the user database and the system will return to normal.
|
||||
Again, the details of your JupyterHub deployment (e.g., whether it's deployed on a scalable cluster)
|
||||
will affect the resiliency of the deployment.
|
||||
|
||||
### What interfaces does JupyterHub support?
|
||||
|
||||
Out of the box, JupyterHub supports a variety of popular data science interfaces for user sessions,
|
||||
such as JupyterLab, Jupyter Notebooks, and RStudio. Any interface that can be served
|
||||
via a web address can be served with a JupyterHub (with the right setup).
|
||||
|
||||
### Does JupyterHub make it easier for our team to collaborate?
|
||||
|
||||
JupyterHub provides a standardized environment and access to shared resources for your teams.
|
||||
This greatly reduces the cost associated with sharing analyses and content with other team
|
||||
members, and makes it easier to collaborate and build off of one another's ideas. Combined with
|
||||
access to high-performance computing and data, JupyterHub provides a common resource to
|
||||
amplify your team's ability to prototype their analyses, scale them to larger data, and then
|
||||
share their results with one another.
|
||||
|
||||
JupyterHub also provides a computational framework to share computational narratives between
|
||||
different levels of an organization. For example, data scientists can share Jupyter Notebooks
|
||||
rendered as [voila dashboards](https://voila.readthedocs.io/en/stable/) with those who are not
|
||||
familiar with programming, or create publicly-available interactive analyses to allow others to
|
||||
interact with your work.
|
||||
|
||||
### Can I use JupyterHub with R/RStudio or other languages and environments?
|
||||
|
||||
Yes, Jupyter is a polyglot project, and there are over 40 community-provided kernels for a variety
|
||||
of languages (the most common being Python, Julia, and R). You can also use a JupyterHub to provide
|
||||
access to other interfaces, such as RStudio, that provide their own access to a language kernel.
|
@@ -41,7 +41,7 @@ port.
|
||||
|
||||
## Set the Proxy's REST API communication URL (optional)
|
||||
|
||||
By default, this REST API listens on port 8081 of `localhost` only.
|
||||
By default, this REST API listens on port 8001 of `localhost` only.
|
||||
The Hub service talks to the proxy via a REST API on a secondary port. The
|
||||
API URL can be configured separately and override the default settings.
|
||||
|
||||
|
@@ -80,6 +80,49 @@ To achieve this, simply omit the configuration settings
|
||||
``c.JupyterHub.ssl_key`` and ``c.JupyterHub.ssl_cert``
|
||||
(setting them to ``None`` does not have the same effect, and is an error).
|
||||
|
||||
.. _authentication-token:
|
||||
|
||||
Proxy authentication token
|
||||
--------------------------
|
||||
|
||||
The Hub authenticates its requests to the Proxy using a secret token that
|
||||
the Hub and Proxy agree upon. Note that this applies to the default
|
||||
``ConfigurableHTTPProxy`` implementation. Not all proxy implementations
|
||||
use an auth token.
|
||||
|
||||
The value of this token should be a random string (for example, generated by
|
||||
``openssl rand -hex 32``). You can store it in the configuration file or an
|
||||
environment variable
|
||||
|
||||
Generating and storing token in the configuration file
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
You can set the value in the configuration file, ``jupyterhub_config.py``:
|
||||
|
||||
.. code-block:: python
|
||||
|
||||
c.ConfigurableHTTPProxy.api_token = 'abc123...' # any random string
|
||||
|
||||
Generating and storing as an environment variable
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
You can pass this value of the proxy authentication token to the Hub and Proxy
|
||||
using the ``CONFIGPROXY_AUTH_TOKEN`` environment variable:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
export CONFIGPROXY_AUTH_TOKEN=$(openssl rand -hex 32)
|
||||
|
||||
This environment variable needs to be visible to the Hub and Proxy.
|
||||
|
||||
Default if token is not set
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
If you don't set the Proxy authentication token, the Hub will generate a random
|
||||
key itself, which means that any time you restart the Hub you **must also
|
||||
restart the Proxy**. If the proxy is a subprocess of the Hub, this should happen
|
||||
automatically (this is the default configuration).
|
||||
|
||||
.. _cookie-secret:
|
||||
|
||||
Cookie secret
|
||||
@@ -146,41 +189,73 @@ itself, ``jupyterhub_config.py``, as a binary string:
|
||||
If the cookie secret value changes for the Hub, all single-user notebook
|
||||
servers must also be restarted.
|
||||
|
||||
.. _cookies:
|
||||
|
||||
.. _authentication-token:
|
||||
Cookies used by JupyterHub authentication
|
||||
-----------------------------------------
|
||||
|
||||
Proxy authentication token
|
||||
--------------------------
|
||||
The following cookies are used by the Hub for handling user authentication.
|
||||
|
||||
The Hub authenticates its requests to the Proxy using a secret token that
|
||||
the Hub and Proxy agree upon. The value of this string should be a random
|
||||
string (for example, generated by ``openssl rand -hex 32``).
|
||||
This section was created based on this post_ from Discourse.
|
||||
|
||||
Generating and storing token in the configuration file
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
.. _post: https://discourse.jupyter.org/t/how-to-force-re-login-for-users/1998/6
|
||||
|
||||
Or you can set the value in the configuration file, ``jupyterhub_config.py``:
|
||||
jupyterhub-hub-login
|
||||
~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
.. code-block:: python
|
||||
This is the login token used when visiting Hub-served pages that are
|
||||
protected by authentication such as the main home, the spawn form, etc.
|
||||
If this cookie is set, then the user is logged in.
|
||||
|
||||
c.JupyterHub.proxy_auth_token = '0bc02bede919e99a26de1e2a7a5aadfaf6228de836ec39a05a6c6942831d8fe5'
|
||||
Resetting the Hub cookie secret effectively revokes this cookie.
|
||||
|
||||
Generating and storing as an environment variable
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
This cookie is restricted to the path ``/hub/``.
|
||||
|
||||
You can pass this value of the proxy authentication token to the Hub and Proxy
|
||||
using the ``CONFIGPROXY_AUTH_TOKEN`` environment variable:
|
||||
jupyterhub-user-<username>
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
.. code-block:: bash
|
||||
This is the cookie used for authenticating with a single-user server.
|
||||
It is set by the single-user server after OAuth with the Hub.
|
||||
|
||||
export CONFIGPROXY_AUTH_TOKEN=$(openssl rand -hex 32)
|
||||
Effectively the same as ``jupyterhub-hub-login``, but for the
|
||||
single-user server instead of the Hub. It contains an OAuth access token,
|
||||
which is checked with the Hub to authenticate the browser.
|
||||
|
||||
This environment variable needs to be visible to the Hub and Proxy.
|
||||
Each OAuth access token is associated with a session id (see ``jupyterhub-session-id`` section
|
||||
below).
|
||||
|
||||
Default if token is not set
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
To avoid hitting the Hub on every request, the authentication response
|
||||
is cached. And to avoid a stale cache the cache key is comprised of both
|
||||
the token and session id.
|
||||
|
||||
If you don't set the Proxy authentication token, the Hub will generate a random
|
||||
key itself, which means that any time you restart the Hub you **must also
|
||||
restart the Proxy**. If the proxy is a subprocess of the Hub, this should happen
|
||||
automatically (this is the default configuration).
|
||||
Resetting the Hub cookie secret effectively revokes this cookie.
|
||||
|
||||
This cookie is restricted to the path ``/user/<username>``, so that
|
||||
only the user’s server receives it.
|
||||
|
||||
jupyterhub-session-id
|
||||
~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
This is a random string, meaningless in itself, and the only cookie
|
||||
shared by the Hub and single-user servers.
|
||||
|
||||
Its sole purpose is to coordinate logout of the multiple OAuth cookies.
|
||||
|
||||
This cookie is set to ``/`` so all endpoints can receive it, or clear it, etc.
|
||||
|
||||
jupyterhub-user-<username>-oauth-state
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
A short-lived cookie, used solely to store and validate OAuth state.
|
||||
It is only set while OAuth between the single-user server and the Hub
|
||||
is processing.
|
||||
|
||||
If you use your browser development tools, you should see this cookie
|
||||
for a very brief moment before your are logged in,
|
||||
with an expiration date shorter than ``jupyterhub-hub-login`` or
|
||||
``jupyterhub-user-<username>``.
|
||||
|
||||
This cookie should not exist after you have successfully logged in.
|
||||
|
||||
This cookie is restricted to the path ``/user/<username>``, so that only
|
||||
the user’s server receives it.
|
||||
|
@@ -3,9 +3,9 @@
|
||||
When working with JupyterHub, a **Service** is defined as a process
|
||||
that interacts with the Hub's REST API. A Service may perform a specific
|
||||
or action or task. For example, shutting down individuals' single user
|
||||
notebook servers that have been is a good example of a task that could
|
||||
be automated by a Service. Let's look at how the [cull_idle_servers][]
|
||||
script can be used as a Service.
|
||||
notebook servers that have been idle for some time is a good example of
|
||||
a task that could be automated by a Service. Let's look at how the
|
||||
[cull_idle_servers][] script can be used as a Service.
|
||||
|
||||
## Real-world example to cull idle servers
|
||||
|
||||
|
BIN
docs/source/images/jhub-fluxogram.jpeg
Normal file
BIN
docs/source/images/jhub-fluxogram.jpeg
Normal file
Binary file not shown.
After Width: | Height: | Size: 158 KiB |
15
docs/source/index-about.rst
Normal file
15
docs/source/index-about.rst
Normal file
@@ -0,0 +1,15 @@
|
||||
=====
|
||||
About
|
||||
=====
|
||||
|
||||
JupyterHub is an open source project and community. It is a part of the
|
||||
`Jupyter Project <https://jupyter.org>`_. JupyterHub is an open and inclusive
|
||||
community, and invites contributions from anyone. This section covers information
|
||||
about our community, as well as ways that you can connect and get involved.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
contributor-list
|
||||
changelog
|
||||
gallery-jhub-deployments
|
13
docs/source/index-admin.rst
Normal file
13
docs/source/index-admin.rst
Normal file
@@ -0,0 +1,13 @@
|
||||
=====================
|
||||
Administrator's Guide
|
||||
=====================
|
||||
|
||||
This guide covers best-practices, tips, common questions and operations, as
|
||||
well as other information relevant to running your own JupyterHub over time.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
troubleshooting
|
||||
admin/upgrading
|
||||
changelog
|
@@ -2,21 +2,37 @@
|
||||
JupyterHub
|
||||
==========
|
||||
|
||||
`JupyterHub`_, a multi-user **Hub**, spawns, manages, and proxies multiple
|
||||
`JupyterHub`_ is the best way to serve `Jupyter notebook`_ for multiple users.
|
||||
It can be used in a class of students, a corporate data science group or scientific
|
||||
research group. It is a multi-user **Hub** that spawns, manages, and proxies multiple
|
||||
instances of the single-user `Jupyter notebook`_ server.
|
||||
JupyterHub can be used to serve notebooks to a class of students, a corporate
|
||||
data science group, or a scientific research group.
|
||||
|
||||
.. image:: images/jhub-parts.png
|
||||
To make life easier, JupyterHub has distributions. Be sure to
|
||||
take a look at them before continuing with the configuration of the broad
|
||||
original system of `JupyterHub`_. Today, you can find two main cases:
|
||||
|
||||
1. If you need a simple case for a small amount of users (0-100) and single server
|
||||
take a look at
|
||||
`The Littlest JupyterHub <https://github.com/jupyterhub/the-littlest-jupyterhub>`__ distribution.
|
||||
2. If you need to allow for even more users, a dynamic amount of servers can be used on a cloud,
|
||||
take a look at the `Zero to JupyterHub with Kubernetes <https://github.com/jupyterhub/zero-to-jupyterhub-k8s>`__ .
|
||||
|
||||
|
||||
Four subsystems make up JupyterHub:
|
||||
|
||||
* a **Hub** (tornado process) that is the heart of JupyterHub
|
||||
* a **configurable http proxy** (node-http-proxy) that receives the requests from the client's browser
|
||||
* multiple **single-user Jupyter notebook servers** (Python/IPython/tornado) that are monitored by Spawners
|
||||
* an **authentication class** that manages how users can access the system
|
||||
|
||||
|
||||
Besides these central pieces, you can add optional configurations through a `config.py` file and manage users kernels on an admin panel. A simplification of the whole system can be seen in the figure below:
|
||||
|
||||
.. image:: images/jhub-fluxogram.jpeg
|
||||
:alt: JupyterHub subsystems
|
||||
:width: 40%
|
||||
:align: right
|
||||
:width: 80%
|
||||
:align: center
|
||||
|
||||
Three subsystems make up JupyterHub:
|
||||
|
||||
* a multi-user **Hub** (tornado process)
|
||||
* a **configurable http proxy** (node-http-proxy)
|
||||
* multiple **single-user Jupyter notebook servers** (Python/IPython/tornado)
|
||||
|
||||
JupyterHub performs the following functions:
|
||||
|
||||
@@ -56,46 +72,41 @@ Installation Guide
|
||||
------------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:maxdepth: 2
|
||||
|
||||
installation-guide
|
||||
quickstart
|
||||
quickstart-docker
|
||||
installation-basics
|
||||
|
||||
Getting Started
|
||||
---------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:maxdepth: 2
|
||||
|
||||
getting-started/index
|
||||
getting-started/config-basics
|
||||
getting-started/networking-basics
|
||||
getting-started/security-basics
|
||||
getting-started/authenticators-users-basics
|
||||
getting-started/spawners-basics
|
||||
getting-started/services-basics
|
||||
|
||||
Technical Reference
|
||||
-------------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:maxdepth: 2
|
||||
|
||||
reference/index
|
||||
reference/technical-overview
|
||||
reference/websecurity
|
||||
reference/authenticators
|
||||
reference/spawners
|
||||
reference/services
|
||||
reference/rest
|
||||
reference/templates
|
||||
reference/config-user-env
|
||||
reference/config-examples
|
||||
reference/config-ghoauth
|
||||
reference/config-proxy
|
||||
reference/config-sudo
|
||||
|
||||
Administrators guide
|
||||
--------------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
index-admin
|
||||
|
||||
API Reference
|
||||
-------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
api/index
|
||||
|
||||
Contributing
|
||||
------------
|
||||
@@ -109,51 +120,17 @@ Our `Code of Conduct <https://github.com/jupyter/governance/blob/master/conduct/
|
||||
helps keep our community welcoming to as many people as possible.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:maxdepth: 2
|
||||
|
||||
contributing/community
|
||||
contributing/setup
|
||||
contributing/docs
|
||||
contributing/tests
|
||||
contributing/roadmap
|
||||
contributing/security
|
||||
|
||||
Upgrading JupyterHub
|
||||
--------------------
|
||||
|
||||
We try to make upgrades between minor versions as painless as possible.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
admin/upgrading
|
||||
changelog
|
||||
|
||||
API Reference
|
||||
-------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
api/index
|
||||
|
||||
Troubleshooting
|
||||
---------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
|
||||
troubleshooting
|
||||
contributing/index
|
||||
|
||||
About JupyterHub
|
||||
----------------
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 1
|
||||
:maxdepth: 2
|
||||
|
||||
contributor-list
|
||||
changelog
|
||||
gallery-jhub-deployments
|
||||
index-about
|
||||
|
||||
Indices and tables
|
||||
==================
|
||||
@@ -168,24 +145,6 @@ Questions? Suggestions?
|
||||
- `Jupyter mailing list <https://groups.google.com/forum/#!forum/jupyter>`_
|
||||
- `Jupyter website <https://jupyter.org>`_
|
||||
|
||||
.. _contents:
|
||||
|
||||
Full Table of Contents
|
||||
======================
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
installation-guide
|
||||
getting-started/index
|
||||
reference/index
|
||||
api/index
|
||||
troubleshooting
|
||||
contributor-list
|
||||
gallery-jhub-deployments
|
||||
changelog
|
||||
|
||||
|
||||
.. _JupyterHub: https://github.com/jupyterhub/jupyterhub
|
||||
.. _Jupyter notebook: https://jupyter-notebook.readthedocs.io/en/latest/
|
||||
.. _REST API: http://petstore.swagger.io/?url=https://raw.githubusercontent.com/jupyterhub/jupyterhub/master/docs/rest-api.yml#!/default
|
||||
|
347
docs/source/installation-guide-hard.md
Normal file
347
docs/source/installation-guide-hard.md
Normal file
@@ -0,0 +1,347 @@
|
||||
# Install JupyterHub and JupyterLab from the ground up
|
||||
|
||||
The combination of [JupyterHub](https://jupyterhub.readthedocs.io) and [JupyterLab](https://jupyterlab.readthedocs.io)
|
||||
is a great way to make shared computing resources available to a group.
|
||||
|
||||
These instructions are a guide for a manual, 'bare metal' install of [JupyterHub](https://jupyterhub.readthedocs.io)
|
||||
and [JupyterLab](https://jupyterlab.readthedocs.io). This is ideal for running on a single server: build a beast
|
||||
of a machine and share it within your lab, or use a virtual machine from any VPS or cloud provider.
|
||||
|
||||
This guide has similar goals to [The Littlest JupyterHub](https://the-littlest-jupyterhub.readthedocs.io) setup
|
||||
script. However, instead of bundling all these step for you into one installer, we will perform every step manually.
|
||||
This makes it easy to customize any part (e.g. if you want to run other services on the same system and need to make them
|
||||
work together), as well as giving you full control and understanding of your setup.
|
||||
|
||||
|
||||
## Prerequisites
|
||||
|
||||
Your own server with administrator (root) access. This could be a local machine, a remotely hosted one, or a cloud instance
|
||||
or VPS. Each user who will access JupyterHub should have a standard user account on the machine. The install will be done
|
||||
through the command line - useful if you log into your machine remotely using SSH.
|
||||
|
||||
This tutorial was tested on **Ubuntu 18.04**. No other Linux distributions have been tested, but the instructions
|
||||
should be reasonably straightforward to adapt.
|
||||
|
||||
|
||||
## Goals
|
||||
|
||||
JupyterLab enables access to a multiple 'kernels', each one being a given environment for a given language. The most
|
||||
common is a Python environment, for scientific computing usually one managed by the `conda` package manager.
|
||||
|
||||
This guide will set up JupyterHub and JupyterLab seperately from the Python environment. In other words, we treat
|
||||
JupyterHub+JupyterLab as a 'app' or webservice, which will connect to the kernels available on the system. Specifically:
|
||||
|
||||
- We will create an installation of JupyterHub and JupyterLab using a virtualenv under `/opt` using the system Python.
|
||||
|
||||
- We will install conda globally.
|
||||
|
||||
- We will create a shared conda environment which can be used (but not modified) by all users.
|
||||
|
||||
- We will show how users can create their own private conda environments, where they can install whatever they like.
|
||||
|
||||
|
||||
The default JupyterHub Authenticator uses PAM to authenticate system users with their username and password. One can
|
||||
[choose the authenticator](https://jupyterhub.readthedocs.io/en/stable/reference/authenticators.html#authenticators)
|
||||
that best suits their needs. In this guide we will use the default Authenticator because it makes it easy for everyone to manage data
|
||||
in their home folder and to mix and match different services and access methods (e.g. SSH) which all work using the
|
||||
Linux system user accounts. Therefore, each user of JupyterHub will need a standard system user account.
|
||||
|
||||
Another goal of this guide is to use system provided packages wherever possible. This has the advantage that these packages
|
||||
get automatic patches and security updates (be sure to turn on automatic updates in Ubuntu). This means less maintenance
|
||||
work and a more reliable system.
|
||||
|
||||
## Part 1: JupyterHub and JupyterLab
|
||||
|
||||
### Setup the JupyterHub and JupyterLab in a virtual environment
|
||||
|
||||
First we create a virtual environment under '/opt/jupyterhub'. The '/opt' folder is where apps not belonging to the operating
|
||||
system are [commonly installed](https://unix.stackexchange.com/questions/11544/what-is-the-difference-between-opt-and-usr-local).
|
||||
Both jupyterlab and jupyterhub will be installed into this virtualenv. Create it with the command:
|
||||
|
||||
```sh
|
||||
sudo python3 -m venv /opt/jupyterhub/
|
||||
```
|
||||
|
||||
Now we use pip to install the required Python packages into the new virtual environment. Be sure to install
|
||||
`wheel` first. Since we are separating the user interface from the computing kernels, we don't install
|
||||
any Python scientific packages here. The only exception is `ipywidgets` because this is needed to allow connection
|
||||
between interactive tools running in the kernel and the user interface.
|
||||
|
||||
Note that we use `/opt/jupyterhub/bin/python3 -m pip install` each time - this [makes sure](https://snarky.ca/why-you-should-use-python-m-pip/)
|
||||
that the packages are installed to the correct virtual environment.
|
||||
|
||||
Perform the install using the following commands:
|
||||
|
||||
```sh
|
||||
sudo /opt/jupyterhub/bin/python3 -m pip install wheel
|
||||
sudo /opt/jupyterhub/bin/python3 -m pip install jupyterhub jupyterlab
|
||||
sudo /opt/jupyterhub/bin/python3 -m pip install ipywidgets
|
||||
```
|
||||
|
||||
JupyterHub also currently defaults to requiring `configurable-http-proxy`, which needs `nodejs` and `npm`. The versions
|
||||
of these available in Ubuntu therefore need to be installed first (they are a bit old but this is ok for our needs):
|
||||
|
||||
```sh
|
||||
sudo apt install nodejs npm
|
||||
```
|
||||
|
||||
Then install `configurable-http-proxy`:
|
||||
|
||||
```sh
|
||||
sudo npm install -g configurable-http-proxy
|
||||
```
|
||||
|
||||
### Create the configuration for JupyterHub
|
||||
|
||||
Now we start creating configuration files. To keep everything together, we put all the configuration into the folder
|
||||
created for the virtualenv, under `/opt/jupyterhub/etc/`. For each thing needing configuration, we will create a further
|
||||
subfolder and necessary files.
|
||||
|
||||
First create the folder for the JupyterHub configuration and navigate to it:
|
||||
|
||||
```sh
|
||||
sudo mkdir -p /opt/jupyterhub/etc/jupyterhub/
|
||||
cd /opt/jupyterhub/etc/jupyterhub/
|
||||
```
|
||||
Then generate the default configuration file
|
||||
|
||||
```sh
|
||||
sudo /opt/jupyterhub/bin/jupyterhub --generate-config
|
||||
```
|
||||
This will produce the default configuration file `/opt/jupyterhub/etc/jupyterhub/jupyterhub_config.py`
|
||||
|
||||
You will need to edit the configuration file to make the JupyterLab interface by the default.
|
||||
Set the following configuration option in your `jupyterhub_config.py` file:
|
||||
|
||||
```python
|
||||
c.Spawner.default_url = '/lab'
|
||||
```
|
||||
|
||||
Further configuration options may be found in the documentation.
|
||||
|
||||
### Setup Systemd service
|
||||
|
||||
We will setup JupyterHub to run as a system service using Systemd (which is responsible for managing all services and
|
||||
servers that run on startup in Ubuntu). We will create a service file in a suitable location in the virtualenv folder
|
||||
and then link it to the system services. First create the folder for the service file:
|
||||
|
||||
```sh
|
||||
sudo mkdir -p /opt/jupyterhub/etc/systemd
|
||||
```
|
||||
|
||||
Then create the following text file using your [favourite editor](https://micro-editor.github.io/) at
|
||||
```sh
|
||||
/opt/jupyterhub/etc/systemd/jupyterhub.service
|
||||
```
|
||||
|
||||
Paste the following service unit definition into the file:
|
||||
|
||||
```
|
||||
[Unit]
|
||||
Description=JupyterHub
|
||||
After=syslog.target network.target
|
||||
|
||||
[Service]
|
||||
User=root
|
||||
Environment="PATH=/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/opt/jupyterhub/bin"
|
||||
ExecStart=/opt/jupyterhub/bin/jupyterhub -f /opt/jupyterhub/etc/jupyterhub/jupyterhub_config.py
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
```
|
||||
|
||||
This sets up the environment to use the virtual environment we created, tells Systemd how to start jupyterhub using
|
||||
the configuration file we created, specifies that jupyterhub will be started as the `root` user (needed so that it can
|
||||
start jupyter on behalf of other logged in users), and specifies that jupyterhub should start on boot after the network
|
||||
is enabled.
|
||||
|
||||
Finally, we need to make systemd aware of our service file. First we symlink our file into systemd's directory:
|
||||
|
||||
```sh
|
||||
sudo ln -s /opt/jupyterhub/etc/systemd/jupyterhub.service /etc/systemd/system/jupyterhub.service
|
||||
```
|
||||
|
||||
Then tell systemd to reload its configuration files
|
||||
|
||||
```sh
|
||||
sudo systemctl daemon-reload
|
||||
```
|
||||
|
||||
And finally enable the service
|
||||
|
||||
```sh
|
||||
sudo systemctl enable jupyterhub.service
|
||||
```
|
||||
|
||||
The service will start on reboot, but we can start it straight away using:
|
||||
|
||||
```sh
|
||||
sudo systemctl start jupyterhub.service
|
||||
```
|
||||
|
||||
...and check that it's running using:
|
||||
|
||||
```sh
|
||||
sudo systemctl status jupyterhub.service
|
||||
```
|
||||
|
||||
You should now be already be able to access jupyterhub using `<your servers ip>:8000` (assuming you haven't already set
|
||||
up a firewall or something). However, when you log in the jupyter notebooks will be trying to use the Python virtualenv
|
||||
that was created to install JupyterHub, this is not what we want. So on to part 2
|
||||
|
||||
## Part 2: Conda environments
|
||||
|
||||
### Install conda for the whole system
|
||||
|
||||
We will use `conda` to manage Python environments. We will install the officially maintained `conda` packages for Ubuntu,
|
||||
this means they will get automatic updates with the rest of the system. Setup repo for the official Conda debian packages,
|
||||
instructions are copied from [here](https://docs.conda.io/projects/conda/en/latest/user-guide/install/rpm-debian.html):
|
||||
|
||||
Install Anacononda public gpg key to trusted store
|
||||
```sh
|
||||
curl https://repo.anaconda.com/pkgs/misc/gpgkeys/anaconda.asc | gpg --dearmor > conda.gpg
|
||||
sudo install -o root -g root -m 644 conda.gpg /etc/apt/trusted.gpg.d/
|
||||
```
|
||||
|
||||
Add Debian repo
|
||||
|
||||
```sh
|
||||
echo "deb [arch=amd64] https://repo.anaconda.com/pkgs/misc/debrepo/conda stable main" | sudo tee /etc/apt/sources.list.d/conda.list
|
||||
```
|
||||
|
||||
Install conda
|
||||
|
||||
```sh
|
||||
sudo apt update
|
||||
sudo apt install conda
|
||||
```
|
||||
|
||||
This will install conda into the folder `/opt/conda/`, with the conda command available at `/opt/conda/bin/conda`.
|
||||
|
||||
Finally, we can make conda more easily available to users by symlinking the conda shell setup script to the profile
|
||||
'drop in' folder so that it gets run on login
|
||||
|
||||
```sh
|
||||
sudo ln -s /opt/conda/etc/profile.d/conda.sh /etc/profile.d/conda.sh
|
||||
```
|
||||
|
||||
### Install a default conda environment for all users
|
||||
|
||||
First create a folder for conda envs (might exist already):
|
||||
```sh
|
||||
sudo mkdir /opt/conda/envs/
|
||||
```
|
||||
|
||||
Then create a conda environment to your liking within that folder. Here we have called it 'python' because it will
|
||||
be the obvious default - call it whatever you like. You can install whatever you like into this environment, but you MUST at least install `ipykernel`.
|
||||
|
||||
```sh
|
||||
sudo /opt/conda/bin/conda create --prefix /opt/conda/envs/python python=3.7 ipykernel
|
||||
```
|
||||
|
||||
Once your env is set up as desired, make it visible to Jupyter by installing the kernel spec. There are two options here:
|
||||
|
||||
1 ) Install into the JupyterHub virtualenv - this ensures it overrides the default python version. It will only be visible
|
||||
to the JupyterHub installation we have just created. This is useful to avoid conda environments appearing where they are not expected.
|
||||
|
||||
```sh
|
||||
sudo /opt/conda/envs/python/bin/python -m ipykernel install --prefix=/opt/jupyterhub/ --name 'python' --display-name "Python (default)"
|
||||
```
|
||||
|
||||
2 ) Install it system-wide by putting it into `/usr/local`. It will be visible to any parallel install of JupyterHub or
|
||||
JupyterLab, and will persist even if you later delete or modify the JupyterHub installation. This is useful if the kernels
|
||||
might be used by other services, or if you want to modify the JupyterHub installation independently from the conda environments.
|
||||
|
||||
```sh
|
||||
sudo /opt/conda/envs/python/bin/python -m ipykernel install --prefix /usr/local/ --name 'python' --display-name "Python (default)"
|
||||
````
|
||||
|
||||
### Setting up users' own conda environments
|
||||
|
||||
There is relatively little for the administrator to do here, as users will have to set up their own environments using the shell.
|
||||
On login they should run `conda init` or `/opt/conda/bin/conda`. The can then use conda to set up their environment,
|
||||
although they must also install `ipykernel`. Once done, they can enable their kernel using:
|
||||
|
||||
```sh
|
||||
/path/to/kernel/env/bin/python -m ipykernel install --name 'python-my-env' --display-name "Python My Env"
|
||||
```
|
||||
|
||||
This will place the kernel spec into their home folder, where Jupyter will look for it on startup.
|
||||
|
||||
|
||||
## Setting up a reverse proxy
|
||||
|
||||
The guide so far results in JupyterHub running on port 8000. It is not generally advisable to run open web services in
|
||||
this way - instead, use a reverse proxy running on standard HTTP/HTTPS ports.
|
||||
|
||||
> **Important**: Be aware of the security implications especially if you are running a server that is accessible from the open internet
|
||||
> i.e. not protected within an institutional intranet or private home/office network. You should set up a firewall and
|
||||
> HTTPS encryption, which is outside of the scope of this guide. For HTTPS consider using [LetsEncrypt](https://letsencrypt.org/)
|
||||
> or setting up a [self-signed certificate](https://www.digitalocean.com/community/tutorials/how-to-create-a-self-signed-ssl-certificate-for-nginx-in-ubuntu-18-04).
|
||||
> Firewalls may be set up using `ufw` or `firewalld` and combined with `fail2ban`.
|
||||
|
||||
### Using Nginx
|
||||
Nginx is a mature and established web server and reverse proxy and is easy to install using `sudo apt install nginx`.
|
||||
Details on using Nginx as a reverse proxy can be found elsewhere. Here, we will only outline the additional steps needed
|
||||
to setup JupyterHub with Nginx and host it at a given URL e.g. `<your-server-ip-or-url>/jupyter`.
|
||||
This could be useful for example if you are running several services or web pages on the same server.
|
||||
|
||||
To achieve this needs a few tweaks to both the JupyterHub configuration and the Nginx config. First, edit the
|
||||
configuration file `/opt/jupyterhub/etc/jupyterhub/jupyterhub_config.py` and add the line:
|
||||
|
||||
```python
|
||||
c.JupyterHub.bind_url = 'http://:8000/jupyter'
|
||||
```
|
||||
|
||||
where `/jupyter` will be the relative URL of the JupyterHub.
|
||||
|
||||
Now Nginx must be configured with a to pass all traffic from `/jupyter` to the the local address `127.0.0.1:8000`.
|
||||
Add the following snippet to your nginx configuration file (e.g. `/etc/nginx/sites-available/default`).
|
||||
|
||||
```
|
||||
location /jupyter/ {
|
||||
# NOTE important to also set base url of jupyterhub to /jupyter in its config
|
||||
proxy_pass http://127.0.0.1:8000;
|
||||
|
||||
proxy_redirect off;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Forwarded-Proto $scheme;
|
||||
|
||||
# websocket headers
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection $connection_upgrade;
|
||||
|
||||
}
|
||||
```
|
||||
|
||||
Also add this snippet before the *server* block:
|
||||
|
||||
```
|
||||
map $http_upgrade $connection_upgrade {
|
||||
default upgrade;
|
||||
'' close;
|
||||
}
|
||||
```
|
||||
|
||||
Nginx will not run if there are errors in the configuration, check your configuration using:
|
||||
|
||||
```sh
|
||||
nginx -t
|
||||
```
|
||||
|
||||
If there are no errors, you can restart the Nginx service for the new configuration to take effect.
|
||||
|
||||
```sh
|
||||
sudo systemctl restart nginx.service
|
||||
```
|
||||
|
||||
|
||||
## Getting started using your new JupyterHub
|
||||
|
||||
Once you have setup JupyterHub and Nginx proxy as described, you can browse to your JupyterHub IP or URL
|
||||
(e.g. if your server IP address is `123.456.789.1` and you decided to host JupyterHub at the `/jupyter` URL, browse
|
||||
to `123.456.789.1/jupyter`). You will find a login page where you enter your Linux username and password. On login
|
||||
you will be presented with the JupyterLab interface, with the file browser pane showing the contents of your users'
|
||||
home directory on the server.
|
@@ -1,5 +1,9 @@
|
||||
Installation Guide
|
||||
==================
|
||||
Installation
|
||||
============
|
||||
|
||||
These sections cover how to get up-and-running with JupyterHub. They cover
|
||||
some basics of the tools needed to deploy JupyterHub as well as how to get it
|
||||
running on your own infrastructure.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 3
|
||||
@@ -7,3 +11,4 @@ Installation Guide
|
||||
quickstart
|
||||
quickstart-docker
|
||||
installation-basics
|
||||
installation-guide-hard
|
||||
|
@@ -26,6 +26,10 @@ Before installing JupyterHub, you will need:
|
||||
The `nodejs-legacy` package installs the `node` executable and is currently
|
||||
required for npm to work on Debian/Ubuntu.
|
||||
|
||||
- A [pluggable authentication module (PAM)](https://en.wikipedia.org/wiki/Pluggable_authentication_module)
|
||||
to use the [default Authenticator](./getting-started/authenticators-users-basics.md).
|
||||
PAM is often available by default on most distributions, if this is not the case it can be installed by
|
||||
using the operating system's package manager.
|
||||
- TLS certificate and key for HTTPS communication
|
||||
- Domain name
|
||||
|
||||
|
@@ -52,7 +52,7 @@ c.GitHubOAuthenticator.oauth_callback_url = os.environ['OAUTH_CALLBACK_URL']
|
||||
c.LocalAuthenticator.create_system_users = True
|
||||
|
||||
# specify users and admin
|
||||
c.Authenticator.whitelist = {'rgbkrk', 'minrk', 'jhamrick'}
|
||||
c.Authenticator.allowed_users = {'rgbkrk', 'minrk', 'jhamrick'}
|
||||
c.Authenticator.admin_users = {'jhamrick', 'rgbkrk'}
|
||||
|
||||
# uses the default spawner
|
||||
|
@@ -17,10 +17,12 @@ satisfy the following:
|
||||
Let's start out with needed JupyterHub configuration in `jupyterhub_config.py`:
|
||||
|
||||
```python
|
||||
# Force the proxy to only listen to connections to 127.0.0.1
|
||||
c.JupyterHub.ip = '127.0.0.1'
|
||||
# Force the proxy to only listen to connections to 127.0.0.1 (on port 8000)
|
||||
c.JupyterHub.bind_url = 'http://127.0.0.1:8000'
|
||||
```
|
||||
|
||||
(For Jupyterhub < 0.9 use `c.JupyterHub.ip = '127.0.0.1'`.)
|
||||
|
||||
For high-quality SSL configuration, we also generate Diffie-Helman parameters.
|
||||
This can take a few minutes:
|
||||
|
||||
@@ -81,8 +83,11 @@ server {
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
|
||||
# websocket headers
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection $connection_upgrade;
|
||||
|
||||
proxy_buffering off;
|
||||
}
|
||||
|
||||
# Managing requests to verify letsencrypt host
|
||||
@@ -137,6 +142,20 @@ Now restart `nginx`, restart the JupyterHub, and enjoy accessing
|
||||
`https://HUB.DOMAIN.TLD` while serving other content securely on
|
||||
`https://NO_HUB.DOMAIN.TLD`.
|
||||
|
||||
### SELinux permissions for nginx
|
||||
On distributions with SELinux enabled (e.g. Fedora), one may encounter permission errors
|
||||
when the nginx service is started.
|
||||
|
||||
We need to allow nginx to perform network relay and connect to the jupyterhub port. The
|
||||
following commands do that:
|
||||
|
||||
```bash
|
||||
semanage port -a -t http_port_t -p tcp 8000
|
||||
setsebool -P httpd_can_network_relay 1
|
||||
setsebool -P httpd_can_network_connect 1
|
||||
```
|
||||
Replace 8000 with the port the jupyterhub server is running from.
|
||||
|
||||
|
||||
## Apache
|
||||
|
||||
@@ -197,8 +216,8 @@ In case of the need to run the jupyterhub under /jhub/ or other location please
|
||||
|
||||
httpd.conf amendments:
|
||||
```bash
|
||||
RewriteRule /jhub/(.*) ws://127.0.0.1:8000/jhub/$1 [P,L]
|
||||
RewriteRule /jhub/(.*) http://127.0.0.1:8000/jhub/$1 [P,L]
|
||||
RewriteRule /jhub/(.*) ws://127.0.0.1:8000/jhub/$1 [NE.P,L]
|
||||
RewriteRule /jhub/(.*) http://127.0.0.1:8000/jhub/$1 [NE,P,L]
|
||||
|
||||
ProxyPass /jhub/ http://127.0.0.1:8000/jhub/
|
||||
ProxyPassReverse /jhub/ http://127.0.0.1:8000/jhub/
|
||||
|
30
docs/source/reference/config-reference.rst
Normal file
30
docs/source/reference/config-reference.rst
Normal file
@@ -0,0 +1,30 @@
|
||||
==============================
|
||||
Configuration Reference
|
||||
==============================
|
||||
|
||||
.. important::
|
||||
|
||||
Make sure the version of JupyterHub for this documentation matches your
|
||||
installation version, as the output of this command may change between versions.
|
||||
|
||||
JupyterHub configuration
|
||||
------------------------
|
||||
|
||||
As explained in the `Configuration Basics <../getting-started/config-basics.html#generate-a-default-config-file>`_
|
||||
section, the ``jupyterhub_config.py`` can be automatically generated via
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
jupyterhub --generate-config
|
||||
|
||||
|
||||
The following contains the output of that command for reference.
|
||||
|
||||
.. jupyterhub-generate-config::
|
||||
|
||||
JupyterHub help command output
|
||||
------------------------------
|
||||
|
||||
This section contains the output of the command ``jupyterhub --help-all``.
|
||||
|
||||
.. jupyterhub-help-all::
|
@@ -57,7 +57,7 @@ To do this we add to `/etc/sudoers` (use `visudo` for safe editing of sudoers):
|
||||
For example:
|
||||
|
||||
```bash
|
||||
# comma-separated whitelist of users that can spawn single-user servers
|
||||
# comma-separated list of users that can spawn single-user servers
|
||||
# this should include all of your Hub users
|
||||
Runas_Alias JUPYTER_USERS = rhea, zoe, wash
|
||||
|
||||
@@ -120,6 +120,11 @@ the shadow password database.
|
||||
|
||||
### Shadow group (Linux)
|
||||
|
||||
**Note:** On Fedora based distributions there is no clear way to configure
|
||||
the PAM database to allow sufficient access for authenticating with the target user's password
|
||||
from JupyterHub. As a workaround we recommend use an
|
||||
[alternative authentication method](https://github.com/jupyterhub/jupyterhub/wiki/Authenticators).
|
||||
|
||||
```bash
|
||||
$ ls -l /etc/shadow
|
||||
-rw-r----- 1 root shadow 2197 Jul 21 13:41 shadow
|
||||
|
@@ -1,6 +1,9 @@
|
||||
Technical Reference
|
||||
===================
|
||||
|
||||
This section covers more of the details of the JupyterHub architecture, as well as
|
||||
what happens under-the-hood when you deploy and configure your JupyterHub.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
@@ -13,10 +16,13 @@ Technical Reference
|
||||
proxy
|
||||
separate-proxy
|
||||
rest
|
||||
monitoring
|
||||
database
|
||||
templates
|
||||
../events/index
|
||||
config-user-env
|
||||
config-examples
|
||||
config-ghoauth
|
||||
config-proxy
|
||||
config-sudo
|
||||
config-reference
|
||||
|
20
docs/source/reference/monitoring.rst
Normal file
20
docs/source/reference/monitoring.rst
Normal file
@@ -0,0 +1,20 @@
|
||||
Monitoring
|
||||
==========
|
||||
|
||||
This section covers details on monitoring the state of your JupyterHub installation.
|
||||
|
||||
JupyterHub expose the ``/metrics`` endpoint that returns text describing its current
|
||||
operational state formatted in a way `Prometheus <https://prometheus.io/docs/introduction/overview/>`_ understands.
|
||||
|
||||
Prometheus is a separate open source tool that can be configured to repeatedly poll
|
||||
JupyterHub's ``/metrics`` endpoint to parse and save its current state.
|
||||
|
||||
By doing so, Prometheus can describe JupyterHub's evolving state over time.
|
||||
This evolving state can then be accessed through Prometheus that expose its underlying
|
||||
storage to those allowed to access it, and be presented with dashboards by a
|
||||
tool like `Grafana <https://grafana.com/docs/grafana/latest/getting-started/what-is-grafana/>`_.
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
metrics
|
@@ -19,7 +19,7 @@ In general, for a proxy to be usable by JupyterHub, it must:
|
||||
|
||||
1. support websockets without prior knowledge of the URL where websockets may
|
||||
occur
|
||||
2. support trie-based routing (i.e. allow different routes on `/foo` and
|
||||
2. support trie-based routing (i.e. allow different routes on `/foo` and
|
||||
`/foo/bar` and route based on specificity)
|
||||
3. adding or removing a route should not cause existing connections to drop
|
||||
|
||||
@@ -62,7 +62,7 @@ Hub should call these methods when the Hub itself starts and stops.
|
||||
## Encryption
|
||||
|
||||
When using `internal_ssl` to encrypt traffic behind the proxy, at minimum,
|
||||
your `Proxy` will need client ssl certificates which the `Hub` must be made
|
||||
your `Proxy` will need client ssl certificates which the `Hub` must be made
|
||||
aware of. These can be generated with the command `jupyterhub --generate-certs`
|
||||
which will write them to the `internal_certs_location` in folders named
|
||||
`proxy_api` and `proxy_client`. Alternatively, these can be provided to the
|
||||
@@ -102,7 +102,7 @@ route to be proxied, such as `/user/name/`. A routespec will:
|
||||
### Adding a route
|
||||
|
||||
When adding a route, JupyterHub may pass a JSON-serializable dict as a `data`
|
||||
argument that should be attacked to the proxy route. When that route is
|
||||
argument that should be attached to the proxy route. When that route is
|
||||
retrieved, the `data` argument should be returned as well. If your proxy
|
||||
implementation doesn't support storing data attached to routes, then your
|
||||
Python wrapper may have to handle storing the `data` piece itself, e.g in a
|
||||
@@ -204,7 +204,7 @@ setup(
|
||||
```
|
||||
|
||||
If you have added this metadata to your package,
|
||||
users can select your authenticator with the configuration:
|
||||
users can select your proxy with the configuration:
|
||||
|
||||
```python
|
||||
c.JupyterHub.proxy_class = 'mything'
|
||||
|
@@ -57,6 +57,9 @@ generating an API token is available from the JupyterHub user interface:
|
||||
|
||||
## Add API tokens to the config file
|
||||
|
||||
**This is deprecated. We are in no rush to remove this feature,
|
||||
but please consider if service tokens are right for you.**
|
||||
|
||||
You may also add a dictionary of API tokens and usernames to the hub's
|
||||
configuration file, `jupyterhub_config.py` (note that
|
||||
the **key** is the 'secret-token' while the **value** is the 'username'):
|
||||
@@ -67,6 +70,41 @@ c.JupyterHub.api_tokens = {
|
||||
}
|
||||
```
|
||||
|
||||
### Updating to admin services
|
||||
|
||||
The `api_tokens` configuration has been softly deprecated since the introduction of services.
|
||||
We have no plans to remove it,
|
||||
but users are encouraged to use service configuration instead.
|
||||
|
||||
If you have been using `api_tokens` to create an admin user
|
||||
and a token for that user to perform some automations,
|
||||
the services mechanism may be a better fit.
|
||||
If you have the following configuration:
|
||||
|
||||
```python
|
||||
c.JupyterHub.admin_users = {"service-admin",}
|
||||
c.JupyterHub.api_tokens = {
|
||||
"secret-token": "service-admin",
|
||||
}
|
||||
```
|
||||
|
||||
This can be updated to create an admin service, with the following configuration:
|
||||
|
||||
```python
|
||||
c.JupyterHub.services = [
|
||||
{
|
||||
"name": "service-token",
|
||||
"admin": True,
|
||||
"api_token": "secret-token",
|
||||
},
|
||||
]
|
||||
```
|
||||
|
||||
The token will have the same admin permissions,
|
||||
but there will no longer be a user account created to house it.
|
||||
The main noticeable difference is that there will be no notebook server associated with the account
|
||||
and the service will not show up in the various user list pages and APIs.
|
||||
|
||||
## Make an API request
|
||||
|
||||
To authenticate your requests, pass the API token in the request's
|
||||
|
@@ -151,6 +151,8 @@ c.JupyterHub.services = [
|
||||
{
|
||||
'name': 'my-web-service',
|
||||
'url': 'https://10.0.1.1:1984',
|
||||
# any secret >8 characters, you'll use api_token to
|
||||
# authenticate api requests to the hub from your service
|
||||
'api_token': 'super-secret',
|
||||
}
|
||||
]
|
||||
@@ -249,7 +251,7 @@ prefix = os.environ.get('JUPYTERHUB_SERVICE_PREFIX', '/')
|
||||
|
||||
auth = HubAuth(
|
||||
api_token=os.environ['JUPYTERHUB_API_TOKEN'],
|
||||
cookie_cache_max_age=60,
|
||||
cache_max_age=60,
|
||||
)
|
||||
|
||||
app = Flask(__name__)
|
||||
@@ -313,7 +315,7 @@ class MyHandler(HubAuthenticated, web.RequestHandler):
|
||||
The HubAuth will automatically load the desired configuration from the Service
|
||||
environment variables.
|
||||
|
||||
If you want to limit user access, you can whitelist users through either the
|
||||
If you want to limit user access, you can specify allowed users through either the
|
||||
`.hub_users` attribute or `.hub_groups`. These are sets that check against the
|
||||
username and user group list, respectively. If a user matches neither the user
|
||||
list nor the group list, they will not be allowed access. If both are left
|
||||
@@ -331,7 +333,9 @@ and taking note of the following process:
|
||||
1. retrieve the cookie `jupyterhub-services` from the request.
|
||||
2. Make an API request `GET /hub/api/authorizations/cookie/jupyterhub-services/cookie-value`,
|
||||
where cookie-value is the url-encoded value of the `jupyterhub-services` cookie.
|
||||
This request must be authenticated with a Hub API token in the `Authorization` header.
|
||||
This request must be authenticated with a Hub API token in the `Authorization` header,
|
||||
for example using the `api_token` from your [external service's configuration](#externally-managed-services).
|
||||
|
||||
For example, with [requests][]:
|
||||
|
||||
```python
|
||||
@@ -360,7 +364,7 @@ and taking note of the following process:
|
||||
|
||||
An example of using an Externally-Managed Service and authentication is
|
||||
in [nbviewer README][nbviewer example] section on securing the notebook viewer,
|
||||
and an example of its configuration is found [here](https://github.com/jupyter/nbviewer/blob/master/nbviewer/providers/base.py#L94).
|
||||
and an example of its configuration is found [here](https://github.com/jupyter/nbviewer/blob/ed942b10a52b6259099e2dd687930871dc8aac22/nbviewer/providers/base.py#L95).
|
||||
nbviewer can also be run as a Hub-Managed Service as described [nbviewer README][nbviewer example]
|
||||
section on securing the notebook viewer.
|
||||
|
||||
|
@@ -25,8 +25,10 @@ Some examples include:
|
||||
run without being root, by spawning an intermediate process via `sudo`
|
||||
- [BatchSpawner](https://github.com/jupyterhub/batchspawner) for spawning remote
|
||||
servers using batch systems
|
||||
- [RemoteSpawner](https://github.com/zonca/remotespawner) to spawn notebooks
|
||||
and a remote server and tunnel the port via SSH
|
||||
- [YarnSpawner](https://github.com/jupyterhub/yarnspawner) for spawning notebook
|
||||
servers in YARN containers on a Hadoop cluster
|
||||
- [SSHSpawner](https://github.com/NERSC/sshspawner) to spawn notebooks
|
||||
on a remote server using SSH
|
||||
|
||||
|
||||
## Spawner control methods
|
||||
@@ -72,7 +74,7 @@ It should return `None` if it is still running,
|
||||
and an integer exit status, otherwise.
|
||||
|
||||
For the local process case, `Spawner.poll` uses `os.kill(PID, 0)`
|
||||
to check if the local process is still running.
|
||||
to check if the local process is still running. On Windows, it uses `psutil.pid_exists`.
|
||||
|
||||
### Spawner.stop
|
||||
|
||||
@@ -193,7 +195,7 @@ setup(
|
||||
```
|
||||
|
||||
If you have added this metadata to your package,
|
||||
users can select your authenticator with the configuration:
|
||||
users can select your spawner with the configuration:
|
||||
|
||||
```python
|
||||
c.JupyterHub.spawner_class = 'myservice'
|
||||
|
@@ -70,7 +70,7 @@ To add announcements to be displayed on a page, you have two options:
|
||||
### Announcement Configuration Variables
|
||||
|
||||
If you set the configuration variable `JupyterHub.template_vars =
|
||||
{'announcement': 'some_text}`, the given `some_text` will be placed on
|
||||
{'announcement': 'some_text'}`, the given `some_text` will be placed on
|
||||
the top of all pages. The more specific variables
|
||||
`announcement_login`, `announcement_spawn`, `announcement_home`, and
|
||||
`announcement_logout` are more specific and only show on their
|
||||
|
@@ -76,7 +76,7 @@ resolves the cross-site issues.
|
||||
|
||||
### Disable user config
|
||||
|
||||
If subdomains are not available or not desirable, JupyterHub provides a a
|
||||
If subdomains are not available or not desirable, JupyterHub provides a
|
||||
configuration option `Spawner.disable_user_config`, which can be set to prevent
|
||||
the user-owned configuration files from being loaded. After implementing this
|
||||
option, PATHs and package installation and PATHs are the other things that the
|
||||
@@ -127,3 +127,11 @@ A handy website for testing your deployment is
|
||||
|
||||
|
||||
[configurable-http-proxy]: https://github.com/jupyterhub/configurable-http-proxy
|
||||
|
||||
## Vulnerability reporting
|
||||
|
||||
If you believe you’ve found a security vulnerability in JupyterHub, or any
|
||||
Jupyter project, please report it to
|
||||
[security@ipython.org](mailto:security@iypthon.org). If you prefer to encrypt
|
||||
your security reports, you can use [this PGP public
|
||||
key](https://jupyter-notebook.readthedocs.io/en/stable/_downloads/ipython_security.asc).
|
||||
|
@@ -7,8 +7,8 @@ problem and how to resolve it.
|
||||
[*Behavior*](#behavior)
|
||||
- JupyterHub proxy fails to start
|
||||
- sudospawner fails to run
|
||||
- What is the default behavior when none of the lists (admin, whitelist,
|
||||
group whitelist) are set?
|
||||
- What is the default behavior when none of the lists (admin, allowed,
|
||||
allowed groups) are set?
|
||||
- JupyterHub Docker container not accessible at localhost
|
||||
|
||||
[*Errors*](#errors)
|
||||
@@ -55,14 +55,14 @@ or add:
|
||||
|
||||
to the config file, `jupyterhub_config.py`.
|
||||
|
||||
### What is the default behavior when none of the lists (admin, whitelist, group whitelist) are set?
|
||||
### What is the default behavior when none of the lists (admin, allowed, allowed groups) are set?
|
||||
|
||||
When nothing is given for these lists, there will be no admins, and all users
|
||||
who can authenticate on the system (i.e. all the unix users on the server with
|
||||
a password) will be allowed to start a server. The whitelist lets you limit
|
||||
this to a particular set of users, and the admin_users lets you specify who
|
||||
a password) will be allowed to start a server. The allowed username set lets you limit
|
||||
this to a particular set of users, and admin_users lets you specify who
|
||||
among them may use the admin interface (not necessary, unless you need to do
|
||||
things like inspect other users' servers, or modify the userlist at runtime).
|
||||
things like inspect other users' servers, or modify the user list at runtime).
|
||||
|
||||
### JupyterHub Docker container not accessible at localhost
|
||||
|
||||
@@ -75,6 +75,50 @@ tell Jupyterhub to start at `0.0.0.0` which is visible to everyone. Try this
|
||||
command:
|
||||
`docker run -p 8000:8000 -d --name jupyterhub jupyterhub/jupyterhub jupyterhub --ip 0.0.0.0 --port 8000`
|
||||
|
||||
### How can I kill ports from JupyterHub managed services that have been orphaned?
|
||||
|
||||
I started JupyterHub + nbgrader on the same host without containers. When I try to restart JupyterHub + nbgrader with this configuration, errors appear that the service accounts cannot start because the ports are being used.
|
||||
|
||||
How can I kill the processes that are using these ports?
|
||||
|
||||
Run the following command:
|
||||
|
||||
sudo kill -9 $(sudo lsof -t -i:<service_port>)
|
||||
|
||||
Where `<service_port>` is the port used by the nbgrader course service. This configuration is specified in `jupyterhub_config.py`.
|
||||
|
||||
### Why am I getting a Spawn failed error message?
|
||||
|
||||
After successfully logging in to JupyterHub with a compatible authenticators, I get a 'Spawn failed' error message in the browser. The JupyterHub logs have `jupyterhub KeyError: "getpwnam(): name not found: <my_user_name>`.
|
||||
|
||||
This issue occurs when the authenticator requires a local system user to exist. In these cases, you need to use a spawner
|
||||
that does not require an existing system user account, such as `DockerSpawner` or `KubeSpawner`.
|
||||
|
||||
### How can I run JupyterHub with sudo but use my current env vars and virtualenv location?
|
||||
|
||||
When launching JupyterHub with `sudo jupyterhub` I get import errors and my environment variables don't work.
|
||||
|
||||
When launching services with `sudo ...` the shell won't have the same environment variables or `PATH`s in place. The most direct way to solve this issue is to use the full path to your python environment and add environment variables. For example:
|
||||
|
||||
```bash
|
||||
sudo MY_ENV=abc123 \
|
||||
/home/foo/venv/bin/python3 \
|
||||
/srv/jupyterhub/jupyterhub
|
||||
```
|
||||
|
||||
### How can I view the logs for JupyterHub or the user's Notebook servers when using the DockerSpawner?
|
||||
|
||||
Use `docker logs <container>` where `<container>` is the container name defined within `docker-compose.yml`. For example, to view the logs of the JupyterHub container use:
|
||||
|
||||
docker logs hub
|
||||
|
||||
By default, the user's notebook server is named `jupyter-<username>` where `username` is the user's username within JupyterHub's db. So if you wanted to see the logs for user `foo` you would use:
|
||||
|
||||
docker logs jupyter-foo
|
||||
|
||||
You can also tail logs to view them in real time using the `-f` option:
|
||||
|
||||
docker logs -f hub
|
||||
|
||||
## Errors
|
||||
|
||||
@@ -108,7 +152,7 @@ You should see a similar 200 message, as above, in the Hub log when you first
|
||||
visit your single-user notebook server. If you don't see this message in the log, it
|
||||
may mean that your single-user notebook server isn't connecting to your Hub.
|
||||
|
||||
If you see 403 (forbidden) like this, it's a token problem:
|
||||
If you see 403 (forbidden) like this, it's likely a token problem:
|
||||
|
||||
```
|
||||
403 GET /hub/api/authorizations/cookie/jupyterhub-token-name/[secret] (@10.0.1.4) 4.14ms
|
||||
@@ -152,6 +196,42 @@ After this, when you start your server via JupyterHub, it will build a
|
||||
new container. If this was the underlying cause of the issue, you should see
|
||||
your server again.
|
||||
|
||||
##### Proxy settings (403 GET)
|
||||
|
||||
When your whole JupyterHub sits behind a organization proxy (*not* a reverse proxy like NGINX as part of your setup and *not* the configurable-http-proxy) the environment variables `HTTP_PROXY`, `HTTPS_PROXY`, `http_proxy` and `https_proxy` might be set. This confuses the jupyterhub-singleuser servers: When connecting to the Hub for authorization they connect via the proxy instead of directly connecting to the Hub on localhost. The proxy might deny the request (403 GET). This results in the singleuser server thinking it has a wrong auth token. To circumvent this you should add `<hub_url>,<hub_ip>,localhost,127.0.0.1` to the environment variables `NO_PROXY` and `no_proxy`.
|
||||
|
||||
### Launching Jupyter Notebooks to run as an externally managed JupyterHub service with the `jupyterhub-singleuser` command returns a `JUPYTERHUB_API_TOKEN` error
|
||||
|
||||
[JupyterHub services](https://jupyterhub.readthedocs.io/en/stable/reference/services.html) allow processes to interact with JupyterHub's REST API. Example use-cases include:
|
||||
|
||||
* **Secure Testing**: provide a canonical Jupyter Notebook for testing production data to reduce the number of entry points into production systems.
|
||||
* **Grading Assignments**: provide access to shared Jupyter Notebooks that may be used for management tasks such grading assignments.
|
||||
* **Private Dashboards**: share dashboards with certain group members.
|
||||
|
||||
If possible, try to run the Jupyter Notebook as an externally managed service with one of the provided [jupyter/docker-stacks](https://github.com/jupyter/docker-stacks).
|
||||
|
||||
Standard JupyterHub installations include a [jupyterhub-singleuser](https://github.com/jupyterhub/jupyterhub/blob/9fdab027daa32c9017845572ad9d5ba1722dbc53/setup.py#L116) command which is built from the `jupyterhub.singleuser:main` method. The `jupyterhub-singleuser` command is the default command when JupyterHub launches single-user Jupyter Notebooks. One of the goals of this command is to make sure the version of JupyterHub installed within the Jupyter Notebook coincides with the version of the JupyterHub server itself.
|
||||
|
||||
If you launch a Jupyter Notebook with the `jupyterhub-singleuser` command directly from the command line the Jupyter Notebook won't have access to the `JUPYTERHUB_API_TOKEN` and will return:
|
||||
|
||||
```
|
||||
JUPYTERHUB_API_TOKEN env is required to run jupyterhub-singleuser.
|
||||
Did you launch it manually?
|
||||
```
|
||||
|
||||
If you plan on testing `jupyterhub-singleuser` independently from JupyterHub, then you can set the api token environment variable. For example, if were to run the single-user Jupyter Notebook on the host, then:
|
||||
|
||||
export JUPYTERHUB_API_TOKEN=my_secret_token
|
||||
jupyterhub-singleuser
|
||||
|
||||
With a docker container, pass in the environment variable with the run command:
|
||||
|
||||
docker run -d \
|
||||
-p 8888:8888 \
|
||||
-e JUPYTERHUB_API_TOKEN=my_secret_token \
|
||||
jupyter/datascience-notebook:latest
|
||||
|
||||
[This example](https://github.com/jupyterhub/jupyterhub/tree/master/examples/service-notebook/external) demonstrates how to combine the use of the `jupyterhub-singleuser` environment variables when launching a Notebook as an externally managed service.
|
||||
|
||||
## How do I...?
|
||||
|
||||
@@ -193,7 +273,7 @@ where `ssl_cert` is example-chained.crt and ssl_key to your private key.
|
||||
|
||||
Then restart JupyterHub.
|
||||
|
||||
See also [JupyterHub SSL encryption](getting-started.md#ssl-encryption).
|
||||
See also [JupyterHub SSL encryption](./getting-started/security-basics.html#ssl-encryption).
|
||||
|
||||
### Install JupyterHub without a network connection
|
||||
|
||||
@@ -252,8 +332,7 @@ notebook servers to default to JupyterLab:
|
||||
### How do I set up JupyterHub for a workshop (when users are not known ahead of time)?
|
||||
|
||||
1. Set up JupyterHub using OAuthenticator for GitHub authentication
|
||||
2. Configure whitelist to be an empty list in` jupyterhub_config.py`
|
||||
3. Configure admin list to have workshop leaders be listed with administrator privileges.
|
||||
2. Configure admin list to have workshop leaders be listed with administrator privileges.
|
||||
|
||||
Users will need a GitHub account to login and be authenticated by the Hub.
|
||||
|
||||
@@ -281,7 +360,6 @@ Or use syslog:
|
||||
|
||||
jupyterhub | logger -t jupyterhub
|
||||
|
||||
|
||||
## Troubleshooting commands
|
||||
|
||||
The following commands provide additional detail about installed packages,
|
||||
|
@@ -1,56 +0,0 @@
|
||||
"""autodoc extension for configurable traits"""
|
||||
from sphinx.domains.python import PyClassmember
|
||||
from sphinx.ext.autodoc import AttributeDocumenter
|
||||
from sphinx.ext.autodoc import ClassDocumenter
|
||||
from traitlets import TraitType
|
||||
from traitlets import Undefined
|
||||
|
||||
|
||||
class ConfigurableDocumenter(ClassDocumenter):
|
||||
"""Specialized Documenter subclass for traits with config=True"""
|
||||
|
||||
objtype = 'configurable'
|
||||
directivetype = 'class'
|
||||
|
||||
def get_object_members(self, want_all):
|
||||
"""Add traits with .tag(config=True) to members list"""
|
||||
check, members = super().get_object_members(want_all)
|
||||
get_traits = (
|
||||
self.object.class_own_traits
|
||||
if self.options.inherited_members
|
||||
else self.object.class_traits
|
||||
)
|
||||
trait_members = []
|
||||
for name, trait in sorted(get_traits(config=True).items()):
|
||||
# put help in __doc__ where autodoc will look for it
|
||||
trait.__doc__ = trait.help
|
||||
trait_members.append((name, trait))
|
||||
return check, trait_members + members
|
||||
|
||||
|
||||
class TraitDocumenter(AttributeDocumenter):
|
||||
objtype = 'trait'
|
||||
directivetype = 'attribute'
|
||||
member_order = 1
|
||||
priority = 100
|
||||
|
||||
@classmethod
|
||||
def can_document_member(cls, member, membername, isattr, parent):
|
||||
return isinstance(member, TraitType)
|
||||
|
||||
def format_name(self):
|
||||
return 'config c.' + super().format_name()
|
||||
|
||||
def add_directive_header(self, sig):
|
||||
default = self.object.get_default_value()
|
||||
if default is Undefined:
|
||||
default_s = ''
|
||||
else:
|
||||
default_s = repr(default)
|
||||
sig = ' = {}({})'.format(self.object.__class__.__name__, default_s)
|
||||
return super().add_directive_header(sig)
|
||||
|
||||
|
||||
def setup(app):
|
||||
app.add_autodocumenter(ConfigurableDocumenter)
|
||||
app.add_autodocumenter(TraitDocumenter)
|
@@ -1,41 +1,4 @@
|
||||
# `cull-idle` Example
|
||||
# idle-culler example
|
||||
|
||||
The `cull_idle_servers.py` file provides a script to cull and shut down idle
|
||||
single-user notebook servers. This script is used when `cull-idle` is run as
|
||||
a Service or when it is run manually as a standalone script.
|
||||
|
||||
|
||||
## Configure `cull-idle` to run as a Hub-Managed Service
|
||||
|
||||
In `jupyterhub_config.py`, add the following dictionary for the `cull-idle`
|
||||
Service to the `c.JupyterHub.services` list:
|
||||
|
||||
```python
|
||||
c.JupyterHub.services = [
|
||||
{
|
||||
'name': 'cull-idle',
|
||||
'admin': True,
|
||||
'command': [sys.executable, 'cull_idle_servers.py', '--timeout=3600'],
|
||||
}
|
||||
]
|
||||
```
|
||||
|
||||
where:
|
||||
|
||||
- `'admin': True` indicates that the Service has 'admin' permissions, and
|
||||
- `'command'` indicates that the Service will be managed by the Hub.
|
||||
|
||||
## Run `cull-idle` manually as a standalone script
|
||||
|
||||
This will run `cull-idle` manually. `cull-idle` can be run as a standalone
|
||||
script anywhere with access to the Hub, and will periodically check for idle
|
||||
servers and shut them down via the Hub's REST API. In order to shutdown the
|
||||
servers, the token given to cull-idle must have admin privileges.
|
||||
|
||||
Generate an API token and store it in the `JUPYTERHUB_API_TOKEN` environment
|
||||
variable. Run `cull_idle_servers.py` manually.
|
||||
|
||||
```bash
|
||||
export JUPYTERHUB_API_TOKEN=$(jupyterhub token)
|
||||
python3 cull_idle_servers.py [--timeout=900] [--url=http://127.0.0.1:8081/hub/api]
|
||||
```
|
||||
The idle culler has been moved to its own repository at
|
||||
[jupyterhub/jupyterhub-idle-culler](https://github.com/jupyterhub/jupyterhub-idle-culler).
|
||||
|
@@ -1,385 +0,0 @@
|
||||
#!/usr/bin/env python3
|
||||
"""script to monitor and cull idle single-user servers
|
||||
|
||||
Caveats:
|
||||
|
||||
last_activity is not updated with high frequency,
|
||||
so cull timeout should be greater than the sum of:
|
||||
|
||||
- single-user websocket ping interval (default: 30s)
|
||||
- JupyterHub.last_activity_interval (default: 5 minutes)
|
||||
|
||||
You can run this as a service managed by JupyterHub with this in your config::
|
||||
|
||||
|
||||
c.JupyterHub.services = [
|
||||
{
|
||||
'name': 'cull-idle',
|
||||
'admin': True,
|
||||
'command': [sys.executable, 'cull_idle_servers.py', '--timeout=3600'],
|
||||
}
|
||||
]
|
||||
|
||||
Or run it manually by generating an API token and storing it in `JUPYTERHUB_API_TOKEN`:
|
||||
|
||||
export JUPYTERHUB_API_TOKEN=$(jupyterhub token)
|
||||
python3 cull_idle_servers.py [--timeout=900] [--url=http://127.0.0.1:8081/hub/api]
|
||||
|
||||
This script uses the same ``--timeout`` and ``--max-age`` values for
|
||||
culling users and users' servers. If you want a different value for
|
||||
users and servers, you should add this script to the services list
|
||||
twice, just with different ``name``s, different values, and one with
|
||||
the ``--cull-users`` option.
|
||||
"""
|
||||
import json
|
||||
import os
|
||||
from datetime import datetime
|
||||
from datetime import timezone
|
||||
from functools import partial
|
||||
|
||||
try:
|
||||
from urllib.parse import quote
|
||||
except ImportError:
|
||||
from urllib import quote
|
||||
|
||||
import dateutil.parser
|
||||
|
||||
from tornado.gen import coroutine, multi
|
||||
from tornado.locks import Semaphore
|
||||
from tornado.log import app_log
|
||||
from tornado.httpclient import AsyncHTTPClient, HTTPRequest
|
||||
from tornado.ioloop import IOLoop, PeriodicCallback
|
||||
from tornado.options import define, options, parse_command_line
|
||||
|
||||
|
||||
def parse_date(date_string):
|
||||
"""Parse a timestamp
|
||||
|
||||
If it doesn't have a timezone, assume utc
|
||||
|
||||
Returned datetime object will always be timezone-aware
|
||||
"""
|
||||
dt = dateutil.parser.parse(date_string)
|
||||
if not dt.tzinfo:
|
||||
# assume naïve timestamps are UTC
|
||||
dt = dt.replace(tzinfo=timezone.utc)
|
||||
return dt
|
||||
|
||||
|
||||
def format_td(td):
|
||||
"""
|
||||
Nicely format a timedelta object
|
||||
|
||||
as HH:MM:SS
|
||||
"""
|
||||
if td is None:
|
||||
return "unknown"
|
||||
if isinstance(td, str):
|
||||
return td
|
||||
seconds = int(td.total_seconds())
|
||||
h = seconds // 3600
|
||||
seconds = seconds % 3600
|
||||
m = seconds // 60
|
||||
seconds = seconds % 60
|
||||
return "{h:02}:{m:02}:{seconds:02}".format(h=h, m=m, seconds=seconds)
|
||||
|
||||
|
||||
@coroutine
|
||||
def cull_idle(
|
||||
url, api_token, inactive_limit, cull_users=False, max_age=0, concurrency=10
|
||||
):
|
||||
"""Shutdown idle single-user servers
|
||||
|
||||
If cull_users, inactive *users* will be deleted as well.
|
||||
"""
|
||||
auth_header = {'Authorization': 'token %s' % api_token}
|
||||
req = HTTPRequest(url=url + '/users', headers=auth_header)
|
||||
now = datetime.now(timezone.utc)
|
||||
client = AsyncHTTPClient()
|
||||
|
||||
if concurrency:
|
||||
semaphore = Semaphore(concurrency)
|
||||
|
||||
@coroutine
|
||||
def fetch(req):
|
||||
"""client.fetch wrapped in a semaphore to limit concurrency"""
|
||||
yield semaphore.acquire()
|
||||
try:
|
||||
return (yield client.fetch(req))
|
||||
finally:
|
||||
yield semaphore.release()
|
||||
|
||||
else:
|
||||
fetch = client.fetch
|
||||
|
||||
resp = yield fetch(req)
|
||||
users = json.loads(resp.body.decode('utf8', 'replace'))
|
||||
futures = []
|
||||
|
||||
@coroutine
|
||||
def handle_server(user, server_name, server):
|
||||
"""Handle (maybe) culling a single server
|
||||
|
||||
Returns True if server is now stopped (user removable),
|
||||
False otherwise.
|
||||
"""
|
||||
log_name = user['name']
|
||||
if server_name:
|
||||
log_name = '%s/%s' % (user['name'], server_name)
|
||||
if server.get('pending'):
|
||||
app_log.warning(
|
||||
"Not culling server %s with pending %s", log_name, server['pending']
|
||||
)
|
||||
return False
|
||||
|
||||
# jupyterhub < 0.9 defined 'server.url' once the server was ready
|
||||
# as an *implicit* signal that the server was ready.
|
||||
# 0.9 adds a dedicated, explicit 'ready' field.
|
||||
# By current (0.9) definitions, servers that have no pending
|
||||
# events and are not ready shouldn't be in the model,
|
||||
# but let's check just to be safe.
|
||||
|
||||
if not server.get('ready', bool(server['url'])):
|
||||
app_log.warning(
|
||||
"Not culling not-ready not-pending server %s: %s", log_name, server
|
||||
)
|
||||
return False
|
||||
|
||||
if server.get('started'):
|
||||
age = now - parse_date(server['started'])
|
||||
else:
|
||||
# started may be undefined on jupyterhub < 0.9
|
||||
age = None
|
||||
|
||||
# check last activity
|
||||
# last_activity can be None in 0.9
|
||||
if server['last_activity']:
|
||||
inactive = now - parse_date(server['last_activity'])
|
||||
else:
|
||||
# no activity yet, use start date
|
||||
# last_activity may be None with jupyterhub 0.9,
|
||||
# which introduces the 'started' field which is never None
|
||||
# for running servers
|
||||
inactive = age
|
||||
|
||||
should_cull = (
|
||||
inactive is not None and inactive.total_seconds() >= inactive_limit
|
||||
)
|
||||
if should_cull:
|
||||
app_log.info(
|
||||
"Culling server %s (inactive for %s)", log_name, format_td(inactive)
|
||||
)
|
||||
|
||||
if max_age and not should_cull:
|
||||
# only check started if max_age is specified
|
||||
# so that we can still be compatible with jupyterhub 0.8
|
||||
# which doesn't define the 'started' field
|
||||
if age is not None and age.total_seconds() >= max_age:
|
||||
app_log.info(
|
||||
"Culling server %s (age: %s, inactive for %s)",
|
||||
log_name,
|
||||
format_td(age),
|
||||
format_td(inactive),
|
||||
)
|
||||
should_cull = True
|
||||
|
||||
if not should_cull:
|
||||
app_log.debug(
|
||||
"Not culling server %s (age: %s, inactive for %s)",
|
||||
log_name,
|
||||
format_td(age),
|
||||
format_td(inactive),
|
||||
)
|
||||
return False
|
||||
|
||||
if server_name:
|
||||
# culling a named server
|
||||
delete_url = url + "/users/%s/servers/%s" % (
|
||||
quote(user['name']),
|
||||
quote(server['name']),
|
||||
)
|
||||
else:
|
||||
delete_url = url + '/users/%s/server' % quote(user['name'])
|
||||
|
||||
req = HTTPRequest(url=delete_url, method='DELETE', headers=auth_header)
|
||||
resp = yield fetch(req)
|
||||
if resp.code == 202:
|
||||
app_log.warning("Server %s is slow to stop", log_name)
|
||||
# return False to prevent culling user with pending shutdowns
|
||||
return False
|
||||
return True
|
||||
|
||||
@coroutine
|
||||
def handle_user(user):
|
||||
"""Handle one user.
|
||||
|
||||
Create a list of their servers, and async exec them. Wait for
|
||||
that to be done, and if all servers are stopped, possibly cull
|
||||
the user.
|
||||
"""
|
||||
# shutdown servers first.
|
||||
# Hub doesn't allow deleting users with running servers.
|
||||
# jupyterhub 0.9 always provides a 'servers' model.
|
||||
# 0.8 only does this when named servers are enabled.
|
||||
if 'servers' in user:
|
||||
servers = user['servers']
|
||||
else:
|
||||
# jupyterhub < 0.9 without named servers enabled.
|
||||
# create servers dict with one entry for the default server
|
||||
# from the user model.
|
||||
# only if the server is running.
|
||||
servers = {}
|
||||
if user['server']:
|
||||
servers[''] = {
|
||||
'last_activity': user['last_activity'],
|
||||
'pending': user['pending'],
|
||||
'url': user['server'],
|
||||
}
|
||||
server_futures = [
|
||||
handle_server(user, server_name, server)
|
||||
for server_name, server in servers.items()
|
||||
]
|
||||
results = yield multi(server_futures)
|
||||
if not cull_users:
|
||||
return
|
||||
# some servers are still running, cannot cull users
|
||||
still_alive = len(results) - sum(results)
|
||||
if still_alive:
|
||||
app_log.debug(
|
||||
"Not culling user %s with %i servers still alive",
|
||||
user['name'],
|
||||
still_alive,
|
||||
)
|
||||
return False
|
||||
|
||||
should_cull = False
|
||||
if user.get('created'):
|
||||
age = now - parse_date(user['created'])
|
||||
else:
|
||||
# created may be undefined on jupyterhub < 0.9
|
||||
age = None
|
||||
|
||||
# check last activity
|
||||
# last_activity can be None in 0.9
|
||||
if user['last_activity']:
|
||||
inactive = now - parse_date(user['last_activity'])
|
||||
else:
|
||||
# no activity yet, use start date
|
||||
# last_activity may be None with jupyterhub 0.9,
|
||||
# which introduces the 'created' field which is never None
|
||||
inactive = age
|
||||
|
||||
should_cull = (
|
||||
inactive is not None and inactive.total_seconds() >= inactive_limit
|
||||
)
|
||||
if should_cull:
|
||||
app_log.info("Culling user %s (inactive for %s)", user['name'], inactive)
|
||||
|
||||
if max_age and not should_cull:
|
||||
# only check created if max_age is specified
|
||||
# so that we can still be compatible with jupyterhub 0.8
|
||||
# which doesn't define the 'started' field
|
||||
if age is not None and age.total_seconds() >= max_age:
|
||||
app_log.info(
|
||||
"Culling user %s (age: %s, inactive for %s)",
|
||||
user['name'],
|
||||
format_td(age),
|
||||
format_td(inactive),
|
||||
)
|
||||
should_cull = True
|
||||
|
||||
if not should_cull:
|
||||
app_log.debug(
|
||||
"Not culling user %s (created: %s, last active: %s)",
|
||||
user['name'],
|
||||
format_td(age),
|
||||
format_td(inactive),
|
||||
)
|
||||
return False
|
||||
|
||||
req = HTTPRequest(
|
||||
url=url + '/users/%s' % user['name'], method='DELETE', headers=auth_header
|
||||
)
|
||||
yield fetch(req)
|
||||
return True
|
||||
|
||||
for user in users:
|
||||
futures.append((user['name'], handle_user(user)))
|
||||
|
||||
for (name, f) in futures:
|
||||
try:
|
||||
result = yield f
|
||||
except Exception:
|
||||
app_log.exception("Error processing %s", name)
|
||||
else:
|
||||
if result:
|
||||
app_log.debug("Finished culling %s", name)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
define(
|
||||
'url',
|
||||
default=os.environ.get('JUPYTERHUB_API_URL'),
|
||||
help="The JupyterHub API URL",
|
||||
)
|
||||
define('timeout', default=600, help="The idle timeout (in seconds)")
|
||||
define(
|
||||
'cull_every',
|
||||
default=0,
|
||||
help="The interval (in seconds) for checking for idle servers to cull",
|
||||
)
|
||||
define(
|
||||
'max_age',
|
||||
default=0,
|
||||
help="The maximum age (in seconds) of servers that should be culled even if they are active",
|
||||
)
|
||||
define(
|
||||
'cull_users',
|
||||
default=False,
|
||||
help="""Cull users in addition to servers.
|
||||
This is for use in temporary-user cases such as tmpnb.""",
|
||||
)
|
||||
define(
|
||||
'concurrency',
|
||||
default=10,
|
||||
help="""Limit the number of concurrent requests made to the Hub.
|
||||
|
||||
Deleting a lot of users at the same time can slow down the Hub,
|
||||
so limit the number of API requests we have outstanding at any given time.
|
||||
""",
|
||||
)
|
||||
|
||||
parse_command_line()
|
||||
if not options.cull_every:
|
||||
options.cull_every = options.timeout // 2
|
||||
api_token = os.environ['JUPYTERHUB_API_TOKEN']
|
||||
|
||||
try:
|
||||
AsyncHTTPClient.configure("tornado.curl_httpclient.CurlAsyncHTTPClient")
|
||||
except ImportError as e:
|
||||
app_log.warning(
|
||||
"Could not load pycurl: %s\n"
|
||||
"pycurl is recommended if you have a large number of users.",
|
||||
e,
|
||||
)
|
||||
|
||||
loop = IOLoop.current()
|
||||
cull = partial(
|
||||
cull_idle,
|
||||
url=options.url,
|
||||
api_token=api_token,
|
||||
inactive_limit=options.timeout,
|
||||
cull_users=options.cull_users,
|
||||
max_age=options.max_age,
|
||||
concurrency=options.concurrency,
|
||||
)
|
||||
# schedule first cull immediately
|
||||
# because PeriodicCallback doesn't start until the end of the first interval
|
||||
loop.add_callback(cull)
|
||||
# schedule periodic cull
|
||||
pc = PeriodicCallback(cull, 1e3 * options.cull_every)
|
||||
pc.start()
|
||||
try:
|
||||
loop.start()
|
||||
except KeyboardInterrupt:
|
||||
pass
|
@@ -1,11 +0,0 @@
|
||||
import sys
|
||||
|
||||
# run cull-idle as a service
|
||||
|
||||
c.JupyterHub.services = [
|
||||
{
|
||||
'name': 'cull-idle',
|
||||
'admin': True,
|
||||
'command': [sys.executable, 'cull_idle_servers.py', '--timeout=3600'],
|
||||
}
|
||||
]
|
@@ -5,13 +5,11 @@ so all URLs and requests necessary for OAuth with JupyterHub should be in one pl
|
||||
"""
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
from urllib.parse import urlencode
|
||||
from urllib.parse import urlparse
|
||||
|
||||
from tornado import log
|
||||
from tornado import web
|
||||
from tornado.auth import OAuth2Mixin
|
||||
from tornado.httpclient import AsyncHTTPClient
|
||||
from tornado.httpclient import HTTPRequest
|
||||
from tornado.httputil import url_concat
|
||||
|
@@ -4,7 +4,6 @@ import json
|
||||
import os
|
||||
|
||||
from tornado import escape
|
||||
from tornado import gen
|
||||
from tornado import ioloop
|
||||
from tornado import web
|
||||
|
||||
|
@@ -1,6 +1,3 @@
|
||||
import os
|
||||
import sys
|
||||
|
||||
c.JupyterHub.services = [
|
||||
{
|
||||
'name': 'whoami',
|
||||
|
@@ -1,4 +1,3 @@
|
||||
import os
|
||||
import sys
|
||||
|
||||
c.JupyterHub.services = [
|
||||
|
@@ -6,7 +6,6 @@ showing the user their own info.
|
||||
"""
|
||||
import json
|
||||
import os
|
||||
from getpass import getuser
|
||||
from urllib.parse import urlparse
|
||||
|
||||
from tornado.httpserver import HTTPServer
|
||||
@@ -25,6 +24,7 @@ class WhoAmIHandler(HubOAuthenticated, RequestHandler):
|
||||
# `getuser()` here would mean only the user who started the service
|
||||
# can access the service:
|
||||
|
||||
# from getpass import getuser
|
||||
# hub_users = {getuser()}
|
||||
|
||||
@authenticated
|
||||
|
@@ -4,7 +4,6 @@ This serves `/services/whoami/`, authenticated with the Hub, showing the user th
|
||||
"""
|
||||
import json
|
||||
import os
|
||||
from getpass import getuser
|
||||
from urllib.parse import urlparse
|
||||
|
||||
from tornado.httpserver import HTTPServer
|
||||
@@ -21,6 +20,7 @@ class WhoAmIHandler(HubAuthenticated, RequestHandler):
|
||||
# `getuser()` here would mean only the user who started the service
|
||||
# can access the service:
|
||||
|
||||
# from getpass import getuser
|
||||
# hub_users = {getuser()}
|
||||
|
||||
@authenticated
|
||||
|
@@ -10,10 +10,15 @@ class DemoFormSpawner(LocalProcessSpawner):
|
||||
def _options_form_default(self):
|
||||
default_env = "YOURNAME=%s\n" % self.user.name
|
||||
return """
|
||||
<label for="args">Extra notebook CLI arguments</label>
|
||||
<input name="args" placeholder="e.g. --debug"></input>
|
||||
<label for="env">Environment variables (one per line)</label>
|
||||
<textarea name="env">{env}</textarea>
|
||||
<div class="form-group">
|
||||
<label for="args">Extra notebook CLI arguments</label>
|
||||
<input name="args" class="form-control"
|
||||
placeholder="e.g. --debug"></input>
|
||||
</div>
|
||||
<div class="form-group">
|
||||
<label for="env">Environment variables (one per line)</label>
|
||||
<textarea class="form-control" name="env">{env}</textarea>
|
||||
</div>
|
||||
""".format(
|
||||
env=default_env
|
||||
)
|
||||
|
@@ -1,4 +1,7 @@
|
||||
#!/bin/bash
|
||||
set -exuo pipefail
|
||||
|
||||
# build jupyterhub-onbuild image
|
||||
docker build --build-arg BASE_IMAGE=$DOCKER_REPO:$DOCKER_TAG -t ${DOCKER_REPO}-onbuild:$DOCKER_TAG onbuild
|
||||
# build jupyterhub-demo image
|
||||
docker build --build-arg BASE_IMAGE=${DOCKER_REPO}-onbuild:$DOCKER_TAG -t ${DOCKER_REPO}-demo:$DOCKER_TAG demo-image
|
||||
|
@@ -2,8 +2,11 @@
|
||||
set -exuo pipefail
|
||||
|
||||
export ONBUILD=${DOCKER_REPO}-onbuild
|
||||
export DEMO=${DOCKER_REPO}-demo
|
||||
export REPOS="${DOCKER_REPO} ${ONBUILD} ${DEMO}"
|
||||
# push ONBUILD image
|
||||
docker push $ONBUILD:$DOCKER_TAG
|
||||
docker push $DEMO:$DOCKER_TAG
|
||||
|
||||
function get_hub_version() {
|
||||
rm -f hub_version
|
||||
@@ -20,25 +23,20 @@ function get_hub_version() {
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
get_hub_version
|
||||
|
||||
# when building master, push 0.9.0.dev as well
|
||||
docker tag $DOCKER_REPO:$DOCKER_TAG $DOCKER_REPO:$hub_xyz
|
||||
docker push $DOCKER_REPO:$hub_xyz
|
||||
docker tag $ONBUILD:$DOCKER_TAG $ONBUILD:$hub_xyz
|
||||
docker push $ONBUILD:$hub_xyz
|
||||
for repo in ${REPOS}; do
|
||||
# when building master, push 0.9.0.dev as well
|
||||
docker tag $repo:$DOCKER_TAG $repo:$hub_xyz
|
||||
docker push $repo:$hub_xyz
|
||||
|
||||
# when building 0.9.x, push 0.9 as well
|
||||
docker tag $DOCKER_REPO:$DOCKER_TAG $DOCKER_REPO:$hub_xy
|
||||
docker push $DOCKER_REPO:$hub_xy
|
||||
docker tag $ONBUILD:$DOCKER_TAG $ONBUILD:$hub_xy
|
||||
docker push $ONBUILD:$hub_xyz
|
||||
# when building 0.9.x, push 0.9 as well
|
||||
docker tag $repo:$DOCKER_TAG $repo:$hub_xy
|
||||
docker push $repo:$hub_xy
|
||||
|
||||
# if building a stable release, tag latest as well
|
||||
if [[ "$latest" == "1" ]]; then
|
||||
docker tag $DOCKER_REPO:$DOCKER_TAG $DOCKER_REPO:latest
|
||||
docker push $DOCKER_REPO:latest
|
||||
docker tag $ONBUILD:$DOCKER_TAG $ONBUILD:latest
|
||||
docker push $ONBUILD:latest
|
||||
fi
|
||||
# if building a stable release, tag latest as well
|
||||
if [[ "$latest" == "1" ]]; then
|
||||
docker tag $repo:$DOCKER_TAG $repo:latest
|
||||
docker push $repo:latest
|
||||
fi
|
||||
done
|
||||
|
@@ -4,10 +4,10 @@
|
||||
|
||||
version_info = (
|
||||
1,
|
||||
0,
|
||||
0,
|
||||
"b2", # release (b1, rc1, or "" for final or dev)
|
||||
# "dev", # dev or nothing
|
||||
2,
|
||||
1,
|
||||
# "", # release (b1, rc1, or "" for final or dev)
|
||||
# "dev", # dev or nothing for beta/rc/stable releases
|
||||
)
|
||||
|
||||
# pep 440 version: no dot before beta/rc, but before .dev
|
||||
@@ -18,6 +18,15 @@ version_info = (
|
||||
|
||||
__version__ = ".".join(map(str, version_info[:3])) + ".".join(version_info[3:])
|
||||
|
||||
# Singleton flag to only log the major/minor mismatch warning once per mismatch combo.
|
||||
_version_mismatch_warning_logged = {}
|
||||
|
||||
|
||||
def reset_globals():
|
||||
"""Used to reset globals between test cases."""
|
||||
global _version_mismatch_warning_logged
|
||||
_version_mismatch_warning_logged = {}
|
||||
|
||||
|
||||
def _check_version(hub_version, singleuser_version, log):
|
||||
"""Compare Hub and single-user server versions"""
|
||||
@@ -42,19 +51,27 @@ def _check_version(hub_version, singleuser_version, log):
|
||||
hub_major_minor = V(hub_version).version[:2]
|
||||
singleuser_major_minor = V(singleuser_version).version[:2]
|
||||
extra = ""
|
||||
do_log = True
|
||||
if singleuser_major_minor == hub_major_minor:
|
||||
# patch-level mismatch or lower, log difference at debug-level
|
||||
# because this should be fine
|
||||
log_method = log.debug
|
||||
else:
|
||||
# log warning-level for more significant mismatch, such as 0.8 vs 0.9, etc.
|
||||
log_method = log.warning
|
||||
extra = " This could cause failure to authenticate and result in redirect loops!"
|
||||
log_method(
|
||||
"jupyterhub version %s != jupyterhub-singleuser version %s." + extra,
|
||||
hub_version,
|
||||
singleuser_version,
|
||||
)
|
||||
key = '%s-%s' % (hub_version, singleuser_version)
|
||||
global _version_mismatch_warning_logged
|
||||
if _version_mismatch_warning_logged.get(key):
|
||||
do_log = False # We already logged this warning so don't log it again.
|
||||
else:
|
||||
log_method = log.warning
|
||||
extra = " This could cause failure to authenticate and result in redirect loops!"
|
||||
_version_mismatch_warning_logged[key] = True
|
||||
if do_log:
|
||||
log_method(
|
||||
"jupyterhub version %s != jupyterhub-singleuser version %s." + extra,
|
||||
hub_version,
|
||||
singleuser_version,
|
||||
)
|
||||
else:
|
||||
log.debug(
|
||||
"jupyterhub and jupyterhub-singleuser both on version %s" % hub_version
|
||||
|
@@ -28,7 +28,7 @@ if 'jupyterhub' in sys.modules:
|
||||
alembic_logger.propagate = True
|
||||
alembic_logger.parent = app.log
|
||||
else:
|
||||
fileConfig(config.config_file_name)
|
||||
fileConfig(config.config_file_name, disable_existing_loggers=False)
|
||||
else:
|
||||
fileConfig(config.config_file_name)
|
||||
|
||||
|
@@ -198,14 +198,39 @@ class OAuthAuthorizeHandler(OAuthHandler, BaseHandler):
|
||||
raise
|
||||
self.send_oauth_response(headers, body, status)
|
||||
|
||||
def needs_oauth_confirm(self, user, oauth_client):
|
||||
"""Return whether the given oauth client needs to prompt for access for the given user
|
||||
|
||||
Checks list for oauth clients that don't need confirmation
|
||||
|
||||
(i.e. the user's own server)
|
||||
|
||||
.. versionadded: 1.1
|
||||
"""
|
||||
# get the oauth client ids for the user's own server(s)
|
||||
own_oauth_client_ids = set(
|
||||
spawner.oauth_client_id for spawner in user.spawners.values()
|
||||
)
|
||||
if (
|
||||
# it's the user's own server
|
||||
oauth_client.identifier in own_oauth_client_ids
|
||||
# or it's in the global no-confirm list
|
||||
or oauth_client.identifier
|
||||
in self.settings.get('oauth_no_confirm_list', set())
|
||||
):
|
||||
return False
|
||||
# default: require confirmation
|
||||
return True
|
||||
|
||||
@web.authenticated
|
||||
def get(self):
|
||||
async def get(self):
|
||||
"""GET /oauth/authorization
|
||||
|
||||
Render oauth confirmation page:
|
||||
"Server at ... would like permission to ...".
|
||||
|
||||
Users accessing their own server will skip confirmation.
|
||||
Users accessing their own server or a blessed service
|
||||
will skip confirmation.
|
||||
"""
|
||||
|
||||
uri, http_method, body, headers = self.extract_oauth_params()
|
||||
@@ -215,20 +240,25 @@ class OAuthAuthorizeHandler(OAuthHandler, BaseHandler):
|
||||
)
|
||||
credentials = self.add_credentials(credentials)
|
||||
client = self.oauth_provider.fetch_by_client_id(credentials['client_id'])
|
||||
if client.redirect_uri.startswith(self.current_user.url):
|
||||
if not self.needs_oauth_confirm(self.current_user, client):
|
||||
self.log.debug(
|
||||
"Skipping oauth confirmation for %s accessing %s",
|
||||
self.current_user,
|
||||
client.description,
|
||||
)
|
||||
# access to my own server doesn't require oauth confirmation
|
||||
# this is the pre-1.0 behavior for all oauth
|
||||
self._complete_login(uri, headers, scopes, credentials)
|
||||
return
|
||||
|
||||
# Render oauth 'Authorize application...' page
|
||||
auth_state = await self.current_user.get_auth_state()
|
||||
self.write(
|
||||
self.render_template("oauth.html", scopes=scopes, oauth_client=client)
|
||||
self.render_template(
|
||||
"oauth.html",
|
||||
auth_state=auth_state,
|
||||
scopes=scopes,
|
||||
oauth_client=client,
|
||||
)
|
||||
)
|
||||
|
||||
# Errors that should be shown to the user on the provider website
|
||||
@@ -245,9 +275,26 @@ class OAuthAuthorizeHandler(OAuthHandler, BaseHandler):
|
||||
uri, http_method, body, headers = self.extract_oauth_params()
|
||||
referer = self.request.headers.get('Referer', 'no referer')
|
||||
full_url = self.request.full_url()
|
||||
if referer != full_url:
|
||||
# trim protocol, which cannot be trusted with multiple layers of proxies anyway
|
||||
# Referer is set by browser, but full_url can be modified by proxy layers to appear as http
|
||||
# when it is actually https
|
||||
referer_proto, _, stripped_referer = referer.partition("://")
|
||||
referer_proto = referer_proto.lower()
|
||||
req_proto, _, stripped_full_url = full_url.partition("://")
|
||||
req_proto = req_proto.lower()
|
||||
if referer_proto != req_proto:
|
||||
self.log.warning("Protocol mismatch: %s != %s", referer, full_url)
|
||||
if req_proto == "https":
|
||||
# insecure origin to secure target is not allowed
|
||||
raise web.HTTPError(
|
||||
403, "Not allowing authorization form submitted from insecure page"
|
||||
)
|
||||
if stripped_referer != stripped_full_url:
|
||||
# OAuth post must be made to the URL it came from
|
||||
self.log.error("OAuth POST from %s != %s", referer, full_url)
|
||||
self.log.error("Original OAuth POST from %s != %s", referer, full_url)
|
||||
self.log.error(
|
||||
"Stripped OAuth POST from %s != %s", stripped_referer, stripped_full_url
|
||||
)
|
||||
raise web.HTTPError(
|
||||
403, "Authorization form must be sent from authorization page"
|
||||
)
|
||||
|
@@ -141,6 +141,7 @@ class APIHandler(BaseHandler):
|
||||
'ready': spawner.ready,
|
||||
'state': spawner.get_state() if include_state else None,
|
||||
'url': url_path_join(spawner.user.url, spawner.name, '/'),
|
||||
'user_options': spawner.user_options,
|
||||
'progress_url': spawner._progress_url,
|
||||
}
|
||||
|
||||
|
@@ -3,7 +3,6 @@
|
||||
# Distributed under the terms of the Modified BSD License.
|
||||
import json
|
||||
|
||||
from tornado import gen
|
||||
from tornado import web
|
||||
|
||||
from .. import orm
|
||||
|
@@ -2,12 +2,9 @@
|
||||
# Copyright (c) Jupyter Development Team.
|
||||
# Distributed under the terms of the Modified BSD License.
|
||||
import json
|
||||
from urllib.parse import urlparse
|
||||
|
||||
from tornado import gen
|
||||
from tornado import web
|
||||
|
||||
from .. import orm
|
||||
from ..utils import admin_only
|
||||
from .base import APIHandler
|
||||
|
||||
|
@@ -23,6 +23,7 @@ def service_model(service):
|
||||
'command': service.command,
|
||||
'pid': service.proc.pid if service.proc else 0,
|
||||
'info': service.info,
|
||||
'display': service.display,
|
||||
}
|
||||
|
||||
|
||||
|
@@ -589,11 +589,14 @@ class SpawnProgressAPIHandler(APIHandler):
|
||||
async with aclosing(
|
||||
iterate_until(spawn_future, spawner._generate_progress())
|
||||
) as events:
|
||||
async for event in events:
|
||||
# don't allow events to sneakily set the 'ready' flag
|
||||
if 'ready' in event:
|
||||
event.pop('ready', None)
|
||||
await self.send_event(event)
|
||||
try:
|
||||
async for event in events:
|
||||
# don't allow events to sneakily set the 'ready' flag
|
||||
if 'ready' in event:
|
||||
event.pop('ready', None)
|
||||
await self.send_event(event)
|
||||
except asyncio.CancelledError:
|
||||
pass
|
||||
|
||||
# progress finished, wait for spawn to actually resolve,
|
||||
# in case progress finished early
|
||||
@@ -622,14 +625,14 @@ def _parse_timestamp(timestamp):
|
||||
|
||||
- raise HTTPError(400) on parse error
|
||||
- handle and strip tz info for internal consistency
|
||||
(we use naïve utc timestamps everywhere)
|
||||
(we use naive utc timestamps everywhere)
|
||||
"""
|
||||
try:
|
||||
dt = parse_date(timestamp)
|
||||
except Exception:
|
||||
raise web.HTTPError(400, "Not a valid timestamp: %r", timestamp)
|
||||
if dt.tzinfo:
|
||||
# strip timezone info to naïve UTC datetime
|
||||
# strip timezone info to naive UTC datetime
|
||||
dt = dt.astimezone(timezone.utc).replace(tzinfo=None)
|
||||
|
||||
now = datetime.utcnow()
|
||||
|
@@ -5,17 +5,21 @@
|
||||
import asyncio
|
||||
import atexit
|
||||
import binascii
|
||||
import json
|
||||
import logging
|
||||
import os
|
||||
import re
|
||||
import signal
|
||||
import socket
|
||||
import sys
|
||||
import time
|
||||
from concurrent.futures import ThreadPoolExecutor
|
||||
from datetime import datetime
|
||||
from datetime import timedelta
|
||||
from datetime import timezone
|
||||
from functools import partial
|
||||
from getpass import getuser
|
||||
from glob import glob
|
||||
from operator import itemgetter
|
||||
from textwrap import dedent
|
||||
from urllib.parse import unquote
|
||||
@@ -36,7 +40,6 @@ from tornado.ioloop import IOLoop, PeriodicCallback
|
||||
from tornado.log import app_log, access_log, gen_log
|
||||
import tornado.options
|
||||
from tornado import gen, web
|
||||
from tornado.platform.asyncio import AsyncIOMainLoop
|
||||
|
||||
from traitlets import (
|
||||
Unicode,
|
||||
@@ -52,11 +55,15 @@ from traitlets import (
|
||||
Instance,
|
||||
Bytes,
|
||||
Float,
|
||||
Union,
|
||||
observe,
|
||||
default,
|
||||
validate,
|
||||
)
|
||||
from traitlets.config import Application, Configurable, catch_config_error
|
||||
|
||||
from jupyter_telemetry.eventlog import EventLog
|
||||
|
||||
here = os.path.dirname(__file__)
|
||||
|
||||
import jupyterhub
|
||||
@@ -70,8 +77,9 @@ from .user import UserDict
|
||||
from .oauth.provider import make_provider
|
||||
from ._data import DATA_FILES_PATH
|
||||
from .log import CoroutineLogFormatter, log_request
|
||||
from .pagination import Pagination
|
||||
from .proxy import Proxy, ConfigurableHTTPProxy
|
||||
from .traitlets import URLPrefix, Command, EntryPointType
|
||||
from .traitlets import URLPrefix, Command, EntryPointType, Callable
|
||||
from .utils import (
|
||||
maybe_future,
|
||||
url_path_join,
|
||||
@@ -79,6 +87,10 @@ from .utils import (
|
||||
print_ps_info,
|
||||
make_ssl_context,
|
||||
)
|
||||
from .metrics import HUB_STARTUP_DURATION_SECONDS
|
||||
from .metrics import INIT_SPAWNERS_DURATION_SECONDS
|
||||
from .metrics import RUNNING_SERVERS
|
||||
from .metrics import TOTAL_USERS
|
||||
|
||||
# classes for config
|
||||
from .auth import Authenticator, PAMAuthenticator
|
||||
@@ -145,8 +157,8 @@ flags = {
|
||||
}
|
||||
|
||||
COOKIE_SECRET_BYTES = (
|
||||
32
|
||||
) # the number of bytes to use when generating new cookie secrets
|
||||
32 # the number of bytes to use when generating new cookie secrets
|
||||
)
|
||||
|
||||
HEX_RE = re.compile('^([a-f0-9]{2})+$', re.IGNORECASE)
|
||||
|
||||
@@ -268,7 +280,7 @@ class JupyterHub(Application):
|
||||
|
||||
@default('classes')
|
||||
def _load_classes(self):
|
||||
classes = [Spawner, Authenticator, CryptKeeper]
|
||||
classes = [Spawner, Authenticator, CryptKeeper, Pagination]
|
||||
for name, trait in self.traits(config=True).items():
|
||||
# load entry point groups into configurable class list
|
||||
# so that they show up in config files, etc.
|
||||
@@ -277,7 +289,7 @@ class JupyterHub(Application):
|
||||
try:
|
||||
cls = entry_point.load()
|
||||
except Exception as e:
|
||||
self.log.warning(
|
||||
self.log.debug(
|
||||
"Failed to load %s entrypoint %r: %r",
|
||||
trait.entry_point_group,
|
||||
key,
|
||||
@@ -303,6 +315,19 @@ class JupyterHub(Application):
|
||||
config_file = Unicode('jupyterhub_config.py', help="The config file to load").tag(
|
||||
config=True
|
||||
)
|
||||
|
||||
@validate("config_file")
|
||||
def _validate_config_file(self, proposal):
|
||||
if not self.generate_config and not os.path.isfile(proposal.value):
|
||||
print(
|
||||
"ERROR: Failed to find specified config file: {}".format(
|
||||
proposal.value
|
||||
),
|
||||
file=sys.stderr,
|
||||
)
|
||||
sys.exit(1)
|
||||
return proposal.value
|
||||
|
||||
generate_config = Bool(False, help="Generate default config file").tag(config=True)
|
||||
generate_certs = Bool(False, help="Generate certs used for internal ssl").tag(
|
||||
config=True
|
||||
@@ -325,6 +350,18 @@ class JupyterHub(Application):
|
||||
redirect_to_server = Bool(
|
||||
True, help="Redirect user to server (if running), instead of control panel."
|
||||
).tag(config=True)
|
||||
activity_resolution = Integer(
|
||||
30,
|
||||
help="""
|
||||
Resolution (in seconds) for updating activity
|
||||
|
||||
If activity is registered that is less than activity_resolution seconds
|
||||
more recent than the current value,
|
||||
the new value will be ignored.
|
||||
|
||||
This avoids too many writes to the Hub database.
|
||||
""",
|
||||
).tag(config=True)
|
||||
last_activity_interval = Integer(
|
||||
300, help="Interval (in seconds) at which to update last-activity timestamps."
|
||||
).tag(config=True)
|
||||
@@ -526,10 +563,23 @@ class JupyterHub(Application):
|
||||
def _url_part_changed(self, change):
|
||||
"""propagate deprecated ip/port/base_url config to the bind_url"""
|
||||
urlinfo = urlparse(self.bind_url)
|
||||
urlinfo = urlinfo._replace(netloc='%s:%i' % (self.ip, self.port))
|
||||
if ':' in self.ip:
|
||||
fmt = '[%s]:%i'
|
||||
else:
|
||||
fmt = '%s:%i'
|
||||
urlinfo = urlinfo._replace(netloc=fmt % (self.ip, self.port))
|
||||
urlinfo = urlinfo._replace(path=self.base_url)
|
||||
bind_url = urlunparse(urlinfo)
|
||||
|
||||
# Warn if both bind_url and ip/port/base_url are set
|
||||
if bind_url != self.bind_url:
|
||||
if self.bind_url != self._bind_url_default():
|
||||
self.log.warning(
|
||||
"Both bind_url and ip/port/base_url have been configured. "
|
||||
"JupyterHub.ip, JupyterHub.port, JupyterHub.base_url are"
|
||||
" deprecated in JupyterHub 0.9,"
|
||||
" please use JupyterHub.bind_url instead."
|
||||
)
|
||||
self.bind_url = bind_url
|
||||
|
||||
bind_url = Unicode(
|
||||
@@ -541,6 +591,22 @@ class JupyterHub(Application):
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
@validate('bind_url')
|
||||
def _validate_bind_url(self, proposal):
|
||||
"""ensure protocol field of bind_url matches ssl"""
|
||||
v = proposal['value']
|
||||
proto, sep, rest = v.partition('://')
|
||||
if self.ssl_cert and proto != 'https':
|
||||
return 'https' + sep + rest
|
||||
elif proto != 'http' and not self.ssl_cert:
|
||||
return 'http' + sep + rest
|
||||
return v
|
||||
|
||||
@default('bind_url')
|
||||
def _bind_url_default(self):
|
||||
proto = 'https' if self.ssl_cert else 'http'
|
||||
return proto + '://:8000'
|
||||
|
||||
subdomain_host = Unicode(
|
||||
'',
|
||||
help="""Run single-user servers on subdomains of this host.
|
||||
@@ -578,7 +644,9 @@ class JupyterHub(Application):
|
||||
|
||||
@default('logo_file')
|
||||
def _logo_file_default(self):
|
||||
return os.path.join(self.data_files_path, 'static', 'images', 'jupyter.png')
|
||||
return os.path.join(
|
||||
self.data_files_path, 'static', 'images', 'jupyterhub-80.png'
|
||||
)
|
||||
|
||||
jinja_environment_options = Dict(
|
||||
help="Supply extra arguments that will be passed to Jinja environment."
|
||||
@@ -674,10 +742,10 @@ class JupyterHub(Application):
|
||||
help="""The ip or hostname for proxies and spawners to use
|
||||
for connecting to the Hub.
|
||||
|
||||
Use when the bind address (`hub_ip`) is 0.0.0.0 or otherwise different
|
||||
Use when the bind address (`hub_ip`) is 0.0.0.0, :: or otherwise different
|
||||
from the connect address.
|
||||
|
||||
Default: when `hub_ip` is 0.0.0.0, use `socket.gethostname()`, otherwise use `hub_ip`.
|
||||
Default: when `hub_ip` is 0.0.0.0 or ::, use `socket.gethostname()`, otherwise use `hub_ip`.
|
||||
|
||||
Note: Some spawners or proxy implementations might not support hostnames. Check your
|
||||
spawner or proxy documentation to see if they have extra requirements.
|
||||
@@ -795,14 +863,14 @@ class JupyterHub(Application):
|
||||
|
||||
api_tokens = Dict(
|
||||
Unicode(),
|
||||
help="""PENDING DEPRECATION: consider using service_tokens
|
||||
help="""PENDING DEPRECATION: consider using services
|
||||
|
||||
Dict of token:username to be loaded into the database.
|
||||
|
||||
Allows ahead-of-time generation of API tokens for use by externally managed services,
|
||||
which authenticate as JupyterHub users.
|
||||
|
||||
Consider using service_tokens for general services that talk to the JupyterHub API.
|
||||
Consider using services for general services that talk to the JupyterHub API.
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
@@ -880,6 +948,25 @@ class JupyterHub(Application):
|
||||
def _authenticator_default(self):
|
||||
return self.authenticator_class(parent=self, db=self.db)
|
||||
|
||||
implicit_spawn_seconds = Float(
|
||||
0,
|
||||
help="""Trigger implicit spawns after this many seconds.
|
||||
|
||||
When a user visits a URL for a server that's not running,
|
||||
they are shown a page indicating that the requested server
|
||||
is not running with a button to spawn the server.
|
||||
|
||||
Setting this to a positive value will redirect the user
|
||||
after this many seconds, effectively clicking this button
|
||||
automatically for the users,
|
||||
automatically beginning the spawn process.
|
||||
|
||||
Warning: this can result in errors and surprising behavior
|
||||
when sharing access URLs to actual servers,
|
||||
since the wrong server is likely to be started.
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
allow_named_servers = Bool(
|
||||
False, help="Allow named single-user servers per user"
|
||||
).tag(config=True)
|
||||
@@ -895,6 +982,22 @@ class JupyterHub(Application):
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
default_server_name = Unicode(
|
||||
"",
|
||||
help="If named servers are enabled, default name of server to spawn or open, e.g. by user-redirect.",
|
||||
).tag(config=True)
|
||||
# Ensure that default_server_name doesn't do anything if named servers aren't allowed
|
||||
_default_server_name = Unicode(
|
||||
help="Non-configurable version exposed to JupyterHub."
|
||||
)
|
||||
|
||||
@default('_default_server_name')
|
||||
def _set_default_server_name(self):
|
||||
if self.allow_named_servers:
|
||||
return self.default_server_name
|
||||
else:
|
||||
return ""
|
||||
|
||||
# class for spawning single-user servers
|
||||
spawner_class = EntryPointType(
|
||||
default_value=LocalProcessSpawner,
|
||||
@@ -968,6 +1071,28 @@ class JupyterHub(Application):
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
init_spawners_timeout = Integer(
|
||||
10,
|
||||
help="""
|
||||
Timeout (in seconds) to wait for spawners to initialize
|
||||
|
||||
Checking if spawners are healthy can take a long time
|
||||
if many spawners are active at hub start time.
|
||||
|
||||
If it takes longer than this timeout to check,
|
||||
init_spawner will be left to complete in the background
|
||||
and the http server is allowed to start.
|
||||
|
||||
A timeout of -1 means wait forever,
|
||||
which can mean a slow startup of the Hub
|
||||
but ensures that the Hub is fully consistent by the time it starts responding to requests.
|
||||
This matches the behavior of jupyterhub 1.0.
|
||||
|
||||
.. versionadded: 1.1.0
|
||||
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
db_url = Unicode(
|
||||
'sqlite:///jupyterhub.sqlite',
|
||||
help="url for the database. e.g. `sqlite:///jupyterhub.sqlite`",
|
||||
@@ -1191,12 +1316,42 @@ class JupyterHub(Application):
|
||||
"""
|
||||
).tag(config=True)
|
||||
|
||||
default_url = Unicode(
|
||||
default_url = Union(
|
||||
[Unicode(), Callable()],
|
||||
help="""
|
||||
The default URL for users when they arrive (e.g. when user directs to "/")
|
||||
|
||||
By default, redirects users to their own server.
|
||||
"""
|
||||
|
||||
Can be a Unicode string (e.g. '/hub/home') or a callable based on the handler object:
|
||||
|
||||
::
|
||||
|
||||
def default_url_fn(handler):
|
||||
user = handler.current_user
|
||||
if user and user.admin:
|
||||
return '/hub/admin'
|
||||
return '/hub/home'
|
||||
|
||||
c.JupyterHub.default_url = default_url_fn
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
user_redirect_hook = Callable(
|
||||
None,
|
||||
allow_none=True,
|
||||
help="""
|
||||
Callable to affect behavior of /user-redirect/
|
||||
|
||||
Receives 4 parameters:
|
||||
1. path - URL path that was provided after /user-redirect/
|
||||
2. request - A Tornado HTTPServerRequest representing the current request.
|
||||
3. user - The currently authenticated user.
|
||||
4. base_url - The base_url of the current hub, for relative redirects
|
||||
|
||||
It should return the new URL to redirect to, or None to preserve
|
||||
current behavior.
|
||||
""",
|
||||
).tag(config=True)
|
||||
|
||||
def init_handlers(self):
|
||||
@@ -1549,22 +1704,22 @@ class JupyterHub(Application):
|
||||
# the admin_users config variable will never be used after this point.
|
||||
# only the database values will be referenced.
|
||||
|
||||
whitelist = [
|
||||
allowed_users = [
|
||||
self.authenticator.normalize_username(name)
|
||||
for name in self.authenticator.whitelist
|
||||
for name in self.authenticator.allowed_users
|
||||
]
|
||||
self.authenticator.whitelist = set(whitelist) # force normalization
|
||||
for username in whitelist:
|
||||
self.authenticator.allowed_users = set(allowed_users) # force normalization
|
||||
for username in allowed_users:
|
||||
if not self.authenticator.validate_username(username):
|
||||
raise ValueError("username %r is not valid" % username)
|
||||
|
||||
if not whitelist:
|
||||
if not allowed_users:
|
||||
self.log.info(
|
||||
"Not using whitelist. Any authenticated user will be allowed."
|
||||
"Not using allowed_users. Any authenticated user will be allowed."
|
||||
)
|
||||
|
||||
# add whitelisted users to the db
|
||||
for name in whitelist:
|
||||
# add allowed users to the db
|
||||
for name in allowed_users:
|
||||
user = orm.User.find(db, name)
|
||||
if user is None:
|
||||
user = orm.User(name=name)
|
||||
@@ -1574,13 +1729,16 @@ class JupyterHub(Application):
|
||||
db.commit()
|
||||
|
||||
# Notify authenticator of all users.
|
||||
# This ensures Auth whitelist is up-to-date with the database.
|
||||
# This lets whitelist be used to set up initial list,
|
||||
# but changes to the whitelist can occur in the database,
|
||||
# This ensures Authenticator.allowed_users is up-to-date with the database.
|
||||
# This lets .allowed_users be used to set up initial list,
|
||||
# but changes to the allowed_users set can occur in the database,
|
||||
# and persist across sessions.
|
||||
total_users = 0
|
||||
for user in db.query(orm.User):
|
||||
try:
|
||||
await maybe_future(self.authenticator.add_user(user))
|
||||
f = self.authenticator.add_user(user)
|
||||
if f:
|
||||
await maybe_future(f)
|
||||
except Exception:
|
||||
self.log.exception("Error adding user %s already in db", user.name)
|
||||
if self.authenticator.delete_invalid_users:
|
||||
@@ -1602,6 +1760,7 @@ class JupyterHub(Application):
|
||||
)
|
||||
)
|
||||
else:
|
||||
total_users += 1
|
||||
# handle database upgrades where user.created is undefined.
|
||||
# we don't want to allow user.created to be undefined,
|
||||
# so initialize it to last_activity (if defined) or now.
|
||||
@@ -1609,9 +1768,11 @@ class JupyterHub(Application):
|
||||
user.created = user.last_activity or datetime.utcnow()
|
||||
db.commit()
|
||||
|
||||
# The whitelist set and the users in the db are now the same.
|
||||
# The allowed_users set and the users in the db are now the same.
|
||||
# From this point on, any user changes should be done simultaneously
|
||||
# to the whitelist set and user db, unless the whitelist is empty (all users allowed).
|
||||
# to the allowed_users set and user db, unless the allowed set is empty (all users allowed).
|
||||
|
||||
TOTAL_USERS.set(total_users)
|
||||
|
||||
async def init_groups(self):
|
||||
"""Load predefined groups into the database"""
|
||||
@@ -1624,11 +1785,11 @@ class JupyterHub(Application):
|
||||
for username in usernames:
|
||||
username = self.authenticator.normalize_username(username)
|
||||
if not (
|
||||
await maybe_future(
|
||||
self.authenticator.check_whitelist(username, None)
|
||||
)
|
||||
await maybe_future(self.authenticator.check_allowed(username, None))
|
||||
):
|
||||
raise ValueError("Username %r is not in whitelist" % username)
|
||||
raise ValueError(
|
||||
"Username %r is not in Authenticator.allowed_users" % username
|
||||
)
|
||||
user = orm.User.find(db, name=username)
|
||||
if user is None:
|
||||
if not self.authenticator.validate_username(username):
|
||||
@@ -1652,11 +1813,20 @@ class JupyterHub(Application):
|
||||
if kind == 'user':
|
||||
name = self.authenticator.normalize_username(name)
|
||||
if not (
|
||||
await maybe_future(self.authenticator.check_whitelist(name, None))
|
||||
await maybe_future(self.authenticator.check_allowed(name, None))
|
||||
):
|
||||
raise ValueError("Token name %r is not in whitelist" % name)
|
||||
raise ValueError(
|
||||
"Token user name %r is not in Authenticator.allowed_users"
|
||||
% name
|
||||
)
|
||||
if not self.authenticator.validate_username(name):
|
||||
raise ValueError("Token name %r is not valid" % name)
|
||||
raise ValueError("Token user name %r is not valid" % name)
|
||||
if kind == 'service':
|
||||
if not any(service["name"] == name for service in self.services):
|
||||
self.log.warning(
|
||||
"Warning: service '%s' not in services, creating implicitly. It is recommended to register services using services list."
|
||||
% name
|
||||
)
|
||||
orm_token = orm.APIToken.find(db, token)
|
||||
if orm_token is None:
|
||||
obj = Class.find(db, name)
|
||||
@@ -1689,17 +1859,27 @@ class JupyterHub(Application):
|
||||
# purge expired tokens hourly
|
||||
purge_expired_tokens_interval = 3600
|
||||
|
||||
def purge_expired_tokens(self):
|
||||
"""purge all expiring token objects from the database
|
||||
|
||||
run periodically
|
||||
"""
|
||||
# this should be all the subclasses of Expiring
|
||||
for cls in (orm.APIToken, orm.OAuthAccessToken, orm.OAuthCode):
|
||||
self.log.debug("Purging expired {name}s".format(name=cls.__name__))
|
||||
cls.purge_expired(self.db)
|
||||
|
||||
async def init_api_tokens(self):
|
||||
"""Load predefined API tokens (for services) into database"""
|
||||
await self._add_tokens(self.service_tokens, kind='service')
|
||||
await self._add_tokens(self.api_tokens, kind='user')
|
||||
purge_expired_tokens = partial(orm.APIToken.purge_expired, self.db)
|
||||
purge_expired_tokens()
|
||||
|
||||
self.purge_expired_tokens()
|
||||
# purge expired tokens hourly
|
||||
# we don't need to be prompt about this
|
||||
# because expired tokens cannot be used anyway
|
||||
pc = PeriodicCallback(
|
||||
purge_expired_tokens, 1e3 * self.purge_expired_tokens_interval
|
||||
self.purge_expired_tokens, 1e3 * self.purge_expired_tokens_interval
|
||||
)
|
||||
pc.start()
|
||||
|
||||
@@ -1813,6 +1993,7 @@ class JupyterHub(Application):
|
||||
)
|
||||
|
||||
async def init_spawners(self):
|
||||
self.log.debug("Initializing spawners")
|
||||
db = self.db
|
||||
|
||||
def _user_summary(user):
|
||||
@@ -1903,21 +2084,43 @@ class JupyterHub(Application):
|
||||
else:
|
||||
self.log.debug("%s not running", spawner._log_name)
|
||||
|
||||
spawner._check_pending = False
|
||||
|
||||
# parallelize checks for running Spawners
|
||||
# run query on extant Server objects
|
||||
# so this is O(running servers) not O(total users)
|
||||
# Server objects can be associated with either a Spawner or a Service,
|
||||
# we are only interested in the ones associated with a Spawner
|
||||
check_futures = []
|
||||
for orm_user in db.query(orm.User):
|
||||
user = self.users[orm_user]
|
||||
self.log.debug("Loading state for %s from db", user.name)
|
||||
for name, orm_spawner in user.orm_spawners.items():
|
||||
if orm_spawner.server is not None:
|
||||
# spawner should be running
|
||||
# instantiate Spawner wrapper and check if it's still alive
|
||||
spawner = user.spawners[name]
|
||||
f = asyncio.ensure_future(check_spawner(user, name, spawner))
|
||||
check_futures.append(f)
|
||||
for orm_server in db.query(orm.Server):
|
||||
orm_spawner = orm_server.spawner
|
||||
if not orm_spawner:
|
||||
# check for orphaned Server rows
|
||||
# this shouldn't happen if we've got our sqlachemy right
|
||||
if not orm_server.service:
|
||||
self.log.warning("deleting orphaned server %s", orm_server)
|
||||
self.db.delete(orm_server)
|
||||
self.db.commit()
|
||||
continue
|
||||
# instantiate Spawner wrapper and check if it's still alive
|
||||
# spawner should be running
|
||||
user = self.users[orm_spawner.user]
|
||||
spawner = user.spawners[orm_spawner.name]
|
||||
self.log.debug("Loading state for %s from db", spawner._log_name)
|
||||
# signal that check is pending to avoid race conditions
|
||||
spawner._check_pending = True
|
||||
f = asyncio.ensure_future(check_spawner(user, spawner.name, spawner))
|
||||
check_futures.append(f)
|
||||
|
||||
# it's important that we get here before the first await
|
||||
# so that we know all spawners are instantiated and in the check-pending state
|
||||
|
||||
# await checks after submitting them all
|
||||
await gen.multi(check_futures)
|
||||
if check_futures:
|
||||
self.log.debug(
|
||||
"Awaiting checks for %i possibly-running spawners", len(check_futures)
|
||||
)
|
||||
await gen.multi(check_futures)
|
||||
db.commit()
|
||||
|
||||
# only perform this query if we are going to log it
|
||||
@@ -1925,6 +2128,10 @@ class JupyterHub(Application):
|
||||
user_summaries = map(_user_summary, self.users.values())
|
||||
self.log.debug("Loaded users:\n%s", '\n'.join(user_summaries))
|
||||
|
||||
active_counts = self.users.count_active_users()
|
||||
RUNNING_SERVERS.set(active_counts['active'])
|
||||
return len(check_futures)
|
||||
|
||||
def init_oauth(self):
|
||||
base_url = self.hub.base_url
|
||||
self.oauth_provider = make_provider(
|
||||
@@ -2004,6 +2211,15 @@ class JupyterHub(Application):
|
||||
else:
|
||||
version_hash = datetime.now().strftime("%Y%m%d%H%M%S")
|
||||
|
||||
oauth_no_confirm_list = set()
|
||||
for service in self._service_map.values():
|
||||
if service.oauth_no_confirm:
|
||||
self.log.warning(
|
||||
"Allowing service %s to complete OAuth without confirmation on an authorization web page",
|
||||
service.name,
|
||||
)
|
||||
oauth_no_confirm_list.add(service.oauth_client_id)
|
||||
|
||||
settings = dict(
|
||||
log_function=log_request,
|
||||
config=self.config,
|
||||
@@ -2011,6 +2227,7 @@ class JupyterHub(Application):
|
||||
db=self.db,
|
||||
proxy=self.proxy,
|
||||
hub=self.hub,
|
||||
activity_resolution=self.activity_resolution,
|
||||
admin_users=self.authenticator.admin_users,
|
||||
admin_access=self.admin_access,
|
||||
authenticator=self.authenticator,
|
||||
@@ -2032,9 +2249,12 @@ class JupyterHub(Application):
|
||||
subdomain_host=self.subdomain_host,
|
||||
domain=self.domain,
|
||||
statsd=self.statsd,
|
||||
implicit_spawn_seconds=self.implicit_spawn_seconds,
|
||||
allow_named_servers=self.allow_named_servers,
|
||||
default_server_name=self._default_server_name,
|
||||
named_server_limit_per_user=self.named_server_limit_per_user,
|
||||
oauth_provider=self.oauth_provider,
|
||||
oauth_no_confirm_list=oauth_no_confirm_list,
|
||||
concurrent_spawn_limit=self.concurrent_spawn_limit,
|
||||
spawn_throttle_retry_range=self.spawn_throttle_retry_range,
|
||||
active_server_limit=self.active_server_limit,
|
||||
@@ -2048,6 +2268,8 @@ class JupyterHub(Application):
|
||||
internal_ssl_ca=self.internal_ssl_ca,
|
||||
trusted_alt_names=self.trusted_alt_names,
|
||||
shutdown_on_logout=self.shutdown_on_logout,
|
||||
eventlog=self.eventlog,
|
||||
app=self,
|
||||
)
|
||||
# allow configured settings to have priority
|
||||
settings.update(self.tornado_settings)
|
||||
@@ -2073,6 +2295,16 @@ class JupyterHub(Application):
|
||||
e,
|
||||
)
|
||||
|
||||
def init_eventlog(self):
|
||||
"""Set up the event logging system."""
|
||||
self.eventlog = EventLog(parent=self)
|
||||
|
||||
for dirname, _, files in os.walk(os.path.join(here, 'event-schemas')):
|
||||
for file in files:
|
||||
if not file.endswith('.yaml'):
|
||||
continue
|
||||
self.eventlog.register_schema_file(os.path.join(dirname, file))
|
||||
|
||||
def write_pid_file(self):
|
||||
pid = os.getpid()
|
||||
if self.pid_file:
|
||||
@@ -2082,11 +2314,22 @@ class JupyterHub(Application):
|
||||
|
||||
@catch_config_error
|
||||
async def initialize(self, *args, **kwargs):
|
||||
hub_startup_start_time = time.perf_counter()
|
||||
super().initialize(*args, **kwargs)
|
||||
if self.generate_config or self.generate_certs or self.subapp:
|
||||
return
|
||||
self._start_future = asyncio.Future()
|
||||
|
||||
def record_start(f):
|
||||
startup_time = time.perf_counter() - hub_startup_start_time
|
||||
self.log.debug("It took %.3f seconds for the Hub to start", startup_time)
|
||||
HUB_STARTUP_DURATION_SECONDS.observe(startup_time)
|
||||
|
||||
self._start_future.add_done_callback(record_start)
|
||||
|
||||
self.load_config_file(self.config_file)
|
||||
self.init_logging()
|
||||
self.log.info("Running JupyterHub version %s", jupyterhub.__version__)
|
||||
if 'JupyterHubApp' in self.config:
|
||||
self.log.warning(
|
||||
"Use JupyterHub in config, not JupyterHubApp. Outdated config:\n%s",
|
||||
@@ -2122,7 +2365,9 @@ class JupyterHub(Application):
|
||||
|
||||
_log_cls("Authenticator", self.authenticator_class)
|
||||
_log_cls("Spawner", self.spawner_class)
|
||||
_log_cls("Proxy", self.proxy_class)
|
||||
|
||||
self.init_eventlog()
|
||||
self.init_pycurl()
|
||||
self.init_secrets()
|
||||
self.init_internal_ssl()
|
||||
@@ -2135,11 +2380,62 @@ class JupyterHub(Application):
|
||||
self.init_services()
|
||||
await self.init_api_tokens()
|
||||
self.init_tornado_settings()
|
||||
await self.init_spawners()
|
||||
self.cleanup_oauth_clients()
|
||||
self.init_handlers()
|
||||
self.init_tornado_application()
|
||||
|
||||
# init_spawners can take a while
|
||||
init_spawners_timeout = self.init_spawners_timeout
|
||||
if init_spawners_timeout < 0:
|
||||
# negative timeout means forever (previous, most stable behavior)
|
||||
init_spawners_timeout = 86400
|
||||
|
||||
init_start_time = time.perf_counter()
|
||||
init_spawners_future = asyncio.ensure_future(self.init_spawners())
|
||||
|
||||
def log_init_time(f):
|
||||
n_spawners = f.result()
|
||||
spawner_initialization_time = time.perf_counter() - init_start_time
|
||||
INIT_SPAWNERS_DURATION_SECONDS.observe(spawner_initialization_time)
|
||||
self.log.info(
|
||||
"Initialized %i spawners in %.3f seconds",
|
||||
n_spawners,
|
||||
spawner_initialization_time,
|
||||
)
|
||||
|
||||
init_spawners_future.add_done_callback(log_init_time)
|
||||
|
||||
try:
|
||||
|
||||
# don't allow a zero timeout because we still need to be sure
|
||||
# that the Spawner objects are defined and pending
|
||||
await gen.with_timeout(
|
||||
timedelta(seconds=max(init_spawners_timeout, 1)), init_spawners_future
|
||||
)
|
||||
except gen.TimeoutError:
|
||||
self.log.warning(
|
||||
"init_spawners did not complete within %i seconds. "
|
||||
"Allowing to complete in the background.",
|
||||
self.init_spawners_timeout,
|
||||
)
|
||||
|
||||
if init_spawners_future.done():
|
||||
self.cleanup_oauth_clients()
|
||||
else:
|
||||
# schedule async operations after init_spawners finishes
|
||||
async def finish_init_spawners():
|
||||
await init_spawners_future
|
||||
# schedule cleanup after spawners are all set up
|
||||
# because it relies on the state resolved by init_spawners
|
||||
self.cleanup_oauth_clients()
|
||||
# trigger a proxy check as soon as all spawners are ready
|
||||
# because this may be *after* the check made as part of normal startup.
|
||||
# To avoid races with partially-complete start,
|
||||
# ensure that start is complete before running this check.
|
||||
await self._start_future
|
||||
await self.proxy.check_routes(self.users, self._service_map)
|
||||
|
||||
asyncio.ensure_future(finish_init_spawners())
|
||||
|
||||
async def cleanup(self):
|
||||
"""Shutdown managed services and various subprocesses. Cleanup runtime files."""
|
||||
|
||||
@@ -2184,7 +2480,6 @@ class JupyterHub(Application):
|
||||
self.log.info("Cleaning up PID file %s", self.pid_file)
|
||||
os.remove(self.pid_file)
|
||||
|
||||
# finally stop the loop once we are all cleaned up
|
||||
self.log.info("...done")
|
||||
|
||||
def write_config_file(self):
|
||||
@@ -2248,7 +2543,7 @@ class JupyterHub(Application):
|
||||
continue
|
||||
dt = parse_date(route_data['last_activity'])
|
||||
if dt.tzinfo:
|
||||
# strip timezone info to naïve UTC datetime
|
||||
# strip timezone info to naive UTC datetime
|
||||
dt = dt.astimezone(timezone.utc).replace(tzinfo=None)
|
||||
|
||||
if user.last_activity:
|
||||
@@ -2290,7 +2585,7 @@ class JupyterHub(Application):
|
||||
if self.generate_certs:
|
||||
self.load_config_file(self.config_file)
|
||||
if not self.internal_ssl:
|
||||
self.log.warn(
|
||||
self.log.warning(
|
||||
"You'll need to enable `internal_ssl` "
|
||||
"in the `jupyterhub_config` file to use "
|
||||
"these certs."
|
||||
@@ -2305,6 +2600,20 @@ class JupyterHub(Application):
|
||||
loop.stop()
|
||||
return
|
||||
|
||||
# start the proxy
|
||||
if self.proxy.should_start:
|
||||
try:
|
||||
await self.proxy.start()
|
||||
except Exception as e:
|
||||
self.log.critical("Failed to start proxy", exc_info=True)
|
||||
self.exit(1)
|
||||
else:
|
||||
self.log.info("Not starting proxy")
|
||||
|
||||
# verify that we can talk to the proxy before listening.
|
||||
# avoids delayed failure if we can't talk to the proxy
|
||||
await self.proxy.get_all_routes()
|
||||
|
||||
ssl_context = make_ssl_context(
|
||||
self.internal_ssl_key,
|
||||
self.internal_ssl_cert,
|
||||
@@ -2342,16 +2651,6 @@ class JupyterHub(Application):
|
||||
self.log.error("Failed to bind hub to %s", self.hub.bind_url)
|
||||
raise
|
||||
|
||||
# start the proxy
|
||||
if self.proxy.should_start:
|
||||
try:
|
||||
await self.proxy.start()
|
||||
except Exception as e:
|
||||
self.log.critical("Failed to start proxy", exc_info=True)
|
||||
self.exit(1)
|
||||
else:
|
||||
self.log.info("Not starting proxy")
|
||||
|
||||
# start the service(s)
|
||||
for service_name, service in self._service_map.items():
|
||||
msg = (
|
||||
@@ -2421,25 +2720,76 @@ class JupyterHub(Application):
|
||||
pc.start()
|
||||
|
||||
self.log.info("JupyterHub is now running at %s", self.proxy.public_url)
|
||||
# Use atexit for Windows, it doesn't have signal handling support
|
||||
if _mswindows:
|
||||
atexit.register(self.atexit)
|
||||
# register cleanup on both TERM and INT
|
||||
atexit.register(self.atexit)
|
||||
self.init_signal()
|
||||
self._start_future.set_result(None)
|
||||
|
||||
def init_signal(self):
|
||||
signal.signal(signal.SIGTERM, self.sigterm)
|
||||
if hasattr(signal, 'SIGINFO'):
|
||||
signal.signal(signal.SIGINFO, self.log_status)
|
||||
loop = asyncio.get_event_loop()
|
||||
for s in (signal.SIGTERM, signal.SIGINT):
|
||||
if not _mswindows:
|
||||
loop.add_signal_handler(
|
||||
s, lambda s=s: asyncio.ensure_future(self.shutdown_cancel_tasks(s))
|
||||
)
|
||||
else:
|
||||
signal.signal(s, self.win_shutdown_cancel_tasks)
|
||||
|
||||
def log_status(self, signum, frame):
|
||||
if not _mswindows:
|
||||
infosignals = [signal.SIGUSR1]
|
||||
if hasattr(signal, 'SIGINFO'):
|
||||
infosignals.append(signal.SIGINFO)
|
||||
for s in infosignals:
|
||||
loop.add_signal_handler(
|
||||
s, lambda s=s: asyncio.ensure_future(self.log_status(s))
|
||||
)
|
||||
|
||||
async def log_status(self, sig):
|
||||
"""Log current status, triggered by SIGINFO (^T in many terminals)"""
|
||||
self.log.debug("Received signal %s[%s]", signum, signal.getsignal(signum))
|
||||
self.log.critical("Received signal %s...", sig.name)
|
||||
print_ps_info()
|
||||
print_stacks()
|
||||
|
||||
def sigterm(self, signum, frame):
|
||||
self.log.critical("Received SIGTERM, shutting down")
|
||||
def win_shutdown_cancel_tasks(self, signum, frame):
|
||||
self.log.critical("Received signalnum %s, , initiating shutdown...", signum)
|
||||
raise SystemExit(128 + signum)
|
||||
|
||||
def _init_asyncio_patch(self):
|
||||
"""Set default asyncio policy to be compatible with Tornado.
|
||||
|
||||
Tornado 6 (at least) is not compatible with the default
|
||||
asyncio implementation on Windows.
|
||||
|
||||
Pick the older SelectorEventLoopPolicy on Windows
|
||||
if the known-incompatible default policy is in use.
|
||||
|
||||
Do this as early as possible to make it a low priority and overrideable.
|
||||
|
||||
ref: https://github.com/tornadoweb/tornado/issues/2608
|
||||
|
||||
FIXME: If/when tornado supports the defaults in asyncio,
|
||||
remove and bump tornado requirement for py38.
|
||||
"""
|
||||
if sys.platform.startswith("win") and sys.version_info >= (3, 8):
|
||||
try:
|
||||
from asyncio import (
|
||||
WindowsProactorEventLoopPolicy,
|
||||
WindowsSelectorEventLoopPolicy,
|
||||
)
|
||||
except ImportError:
|
||||
pass
|
||||
# not affected
|
||||
else:
|
||||
if (
|
||||
type(asyncio.get_event_loop_policy())
|
||||
is WindowsProactorEventLoopPolicy
|
||||
):
|
||||
# WindowsProactorEventLoopPolicy is not compatible with Tornado 6.
|
||||
# Fallback to the pre-3.8 default of WindowsSelectorEventLoopPolicy.
|
||||
asyncio.set_event_loop_policy(WindowsSelectorEventLoopPolicy())
|
||||
|
||||
_atexit_ran = False
|
||||
|
||||
def atexit(self):
|
||||
@@ -2447,6 +2797,7 @@ class JupyterHub(Application):
|
||||
if self._atexit_ran:
|
||||
return
|
||||
self._atexit_ran = True
|
||||
self._init_asyncio_patch()
|
||||
# run the cleanup step (in a new loop, because the interrupted one is unclean)
|
||||
asyncio.set_event_loop(asyncio.new_event_loop())
|
||||
IOLoop.clear_current()
|
||||
@@ -2454,6 +2805,30 @@ class JupyterHub(Application):
|
||||
loop.make_current()
|
||||
loop.run_sync(self.cleanup)
|
||||
|
||||
async def shutdown_cancel_tasks(self, sig):
|
||||
"""Cancel all other tasks of the event loop and initiate cleanup"""
|
||||
self.log.critical("Received signal %s, initiating shutdown...", sig.name)
|
||||
tasks = [
|
||||
t for t in asyncio.Task.all_tasks() if t is not asyncio.Task.current_task()
|
||||
]
|
||||
|
||||
if tasks:
|
||||
self.log.debug("Cancelling pending tasks")
|
||||
[t.cancel() for t in tasks]
|
||||
|
||||
try:
|
||||
await asyncio.wait(tasks)
|
||||
except asyncio.CancelledError as e:
|
||||
self.log.debug("Caught Task CancelledError. Ignoring")
|
||||
except StopAsyncIteration as e:
|
||||
self.log.error("Caught StopAsyncIteration Exception", exc_info=True)
|
||||
|
||||
tasks = [t for t in asyncio.Task.all_tasks()]
|
||||
for t in tasks:
|
||||
self.log.debug("Task status: %s", t)
|
||||
await self.cleanup()
|
||||
asyncio.get_event_loop().stop()
|
||||
|
||||
def stop(self):
|
||||
if not self.io_loop:
|
||||
return
|
||||
@@ -2472,13 +2847,19 @@ class JupyterHub(Application):
|
||||
@classmethod
|
||||
def launch_instance(cls, argv=None):
|
||||
self = cls.instance()
|
||||
AsyncIOMainLoop().install()
|
||||
self._init_asyncio_patch()
|
||||
loop = IOLoop.current()
|
||||
loop.add_callback(self.launch_instance_async, argv)
|
||||
task = asyncio.ensure_future(self.launch_instance_async(argv))
|
||||
try:
|
||||
loop.start()
|
||||
except KeyboardInterrupt:
|
||||
print("\nInterrupted")
|
||||
finally:
|
||||
if task.done():
|
||||
# re-raise exceptions in launch_instance_async
|
||||
task.result()
|
||||
loop.stop()
|
||||
loop.close()
|
||||
|
||||
|
||||
NewToken.classes.append(JupyterHub)
|
||||
|
@@ -7,6 +7,7 @@ import re
|
||||
import sys
|
||||
import warnings
|
||||
from concurrent.futures import ThreadPoolExecutor
|
||||
from functools import partial
|
||||
from shutil import which
|
||||
from subprocess import PIPE
|
||||
from subprocess import Popen
|
||||
@@ -100,41 +101,74 @@ class Authenticator(LoggingConfigurable):
|
||||
"""
|
||||
).tag(config=True)
|
||||
|
||||
whitelist = Set(
|
||||
whitelist = Set(help="Deprecated, use `Authenticator.allowed_users`", config=True,)
|
||||
|
||||
allowed_users = Set(
|
||||
help="""
|
||||
Whitelist of usernames that are allowed to log in.
|
||||
Set of usernames that are allowed to log in.
|
||||
|
||||
Use this with supported authenticators to restrict which users can log in. This is an
|
||||
additional whitelist that further restricts users, beyond whatever restrictions the
|
||||
additional list that further restricts users, beyond whatever restrictions the
|
||||
authenticator has in place.
|
||||
|
||||
If empty, does not perform any additional restriction.
|
||||
|
||||
.. versionchanged:: 1.2
|
||||
`Authenticator.whitelist` renamed to `allowed_users`
|
||||
"""
|
||||
).tag(config=True)
|
||||
|
||||
blacklist = Set(
|
||||
blocked_users = Set(
|
||||
help="""
|
||||
Blacklist of usernames that are not allowed to log in.
|
||||
Set of usernames that are not allowed to log in.
|
||||
|
||||
Use this with supported authenticators to restrict which users can not log in. This is an
|
||||
additional blacklist that further restricts users, beyond whatever restrictions the
|
||||
additional block list that further restricts users, beyond whatever restrictions the
|
||||
authenticator has in place.
|
||||
|
||||
If empty, does not perform any additional restriction.
|
||||
|
||||
.. versionadded: 0.9
|
||||
|
||||
.. versionchanged:: 1.2
|
||||
`Authenticator.blacklist` renamed to `blocked_users`
|
||||
"""
|
||||
).tag(config=True)
|
||||
|
||||
@observe('whitelist')
|
||||
def _check_whitelist(self, change):
|
||||
_deprecated_aliases = {
|
||||
"whitelist": ("allowed_users", "1.2"),
|
||||
"blacklist": ("blocked_users", "1.2"),
|
||||
}
|
||||
|
||||
@observe(*list(_deprecated_aliases))
|
||||
def _deprecated_trait(self, change):
|
||||
"""observer for deprecated traits"""
|
||||
old_attr = change.name
|
||||
new_attr, version = self._deprecated_aliases.get(old_attr)
|
||||
new_value = getattr(self, new_attr)
|
||||
if new_value != change.new:
|
||||
# only warn if different
|
||||
# protects backward-compatible config from warnings
|
||||
# if they set the same value under both names
|
||||
self.log.warning(
|
||||
"{cls}.{old} is deprecated in JupyterHub {version}, use {cls}.{new} instead".format(
|
||||
cls=self.__class__.__name__,
|
||||
old=old_attr,
|
||||
new=new_attr,
|
||||
version=version,
|
||||
)
|
||||
)
|
||||
setattr(self, new_attr, change.new)
|
||||
|
||||
@observe('allowed_users')
|
||||
def _check_allowed_users(self, change):
|
||||
short_names = [name for name in change['new'] if len(name) <= 1]
|
||||
if short_names:
|
||||
sorted_names = sorted(short_names)
|
||||
single = ''.join(sorted_names)
|
||||
string_set_typo = "set('%s')" % single
|
||||
self.log.warning(
|
||||
"whitelist contains single-character names: %s; did you mean set([%r]) instead of %s?",
|
||||
"Allowed set contains single-character names: %s; did you mean set([%r]) instead of %s?",
|
||||
sorted_names[:8],
|
||||
single,
|
||||
string_set_typo,
|
||||
@@ -206,6 +240,7 @@ class Authenticator(LoggingConfigurable):
|
||||
|
||||
delete_invalid_users = Bool(
|
||||
False,
|
||||
config=True,
|
||||
help="""Delete any users from the database that do not pass validation
|
||||
|
||||
When JupyterHub starts, `.add_user` will be called
|
||||
@@ -260,39 +295,74 @@ class Authenticator(LoggingConfigurable):
|
||||
|
||||
def __init__(self, **kwargs):
|
||||
super().__init__(**kwargs)
|
||||
for method_name in (
|
||||
'check_whitelist',
|
||||
'check_blacklist',
|
||||
'check_group_whitelist',
|
||||
self._init_deprecated_methods()
|
||||
|
||||
def _init_deprecated_methods(self):
|
||||
# handles deprecated signature *and* name
|
||||
# with correct subclass override priority!
|
||||
for old_name, new_name in (
|
||||
('check_whitelist', 'check_allowed'),
|
||||
('check_blacklist', 'check_blocked_users'),
|
||||
('check_group_whitelist', 'check_allowed_groups'),
|
||||
):
|
||||
original_method = getattr(self, method_name, None)
|
||||
if original_method is None:
|
||||
old_method = getattr(self, old_name, None)
|
||||
if old_method is None:
|
||||
# no such method (check_group_whitelist is optional)
|
||||
continue
|
||||
signature = inspect.signature(original_method)
|
||||
if 'authentication' not in signature.parameters:
|
||||
|
||||
# allow old name to have higher priority
|
||||
# if and only if it's defined in a later subclass
|
||||
# than the new name
|
||||
for cls in self.__class__.mro():
|
||||
has_old_name = old_name in cls.__dict__
|
||||
has_new_name = new_name in cls.__dict__
|
||||
if has_new_name:
|
||||
break
|
||||
if has_old_name and not has_new_name:
|
||||
warnings.warn(
|
||||
"{0}.{1} should be renamed to {0}.{2} for JupyterHub >= 1.2".format(
|
||||
cls.__name__, old_name, new_name
|
||||
),
|
||||
DeprecationWarning,
|
||||
)
|
||||
# use old name instead of new
|
||||
# if old name is overridden in subclass
|
||||
def _new_calls_old(old_name, *args, **kwargs):
|
||||
return getattr(self, old_name)(*args, **kwargs)
|
||||
|
||||
setattr(self, new_name, partial(_new_calls_old, old_name))
|
||||
break
|
||||
|
||||
# deprecate pre-1.0 method signatures
|
||||
signature = inspect.signature(old_method)
|
||||
if 'authentication' not in signature.parameters and not any(
|
||||
param.kind == inspect.Parameter.VAR_KEYWORD
|
||||
for param in signature.parameters.values()
|
||||
):
|
||||
# adapt to pre-1.0 signature for compatibility
|
||||
warnings.warn(
|
||||
"""
|
||||
{0}.{1} does not support the authentication argument,
|
||||
added in JupyterHub 1.0.
|
||||
added in JupyterHub 1.0. and is renamed to {2} in JupyterHub 1.2.
|
||||
|
||||
It should have the signature:
|
||||
|
||||
def {1}(self, username, authentication=None):
|
||||
def {2}(self, username, authentication=None):
|
||||
...
|
||||
|
||||
Adapting for compatibility.
|
||||
""".format(
|
||||
self.__class__.__name__, method_name
|
||||
self.__class__.__name__, old_name, new_name
|
||||
),
|
||||
DeprecationWarning,
|
||||
)
|
||||
|
||||
def wrapped_method(username, authentication=None, **kwargs):
|
||||
def wrapped_method(
|
||||
original_method, username, authentication=None, **kwargs
|
||||
):
|
||||
return original_method(username, **kwargs)
|
||||
|
||||
setattr(self, method_name, wrapped_method)
|
||||
setattr(self, old_name, partial(wrapped_method, old_method))
|
||||
|
||||
async def run_post_auth_hook(self, handler, authentication):
|
||||
"""
|
||||
@@ -326,39 +396,45 @@ class Authenticator(LoggingConfigurable):
|
||||
username = self.username_map.get(username, username)
|
||||
return username
|
||||
|
||||
def check_whitelist(self, username, authentication=None):
|
||||
"""Check if a username is allowed to authenticate based on whitelist configuration
|
||||
def check_allowed(self, username, authentication=None):
|
||||
"""Check if a username is allowed to authenticate based on configuration
|
||||
|
||||
Return True if username is allowed, False otherwise.
|
||||
No whitelist means any username is allowed.
|
||||
No allowed_users set means any username is allowed.
|
||||
|
||||
Names are normalized *before* being checked against the whitelist.
|
||||
Names are normalized *before* being checked against the allowed set.
|
||||
|
||||
.. versionchanged:: 1.0
|
||||
Signature updated to accept authentication data and any future changes
|
||||
"""
|
||||
if not self.whitelist:
|
||||
# No whitelist means any name is allowed
|
||||
return True
|
||||
return username in self.whitelist
|
||||
|
||||
def check_blacklist(self, username, authentication=None):
|
||||
"""Check if a username is blocked to authenticate based on blacklist configuration
|
||||
.. versionchanged:: 1.2
|
||||
Renamed check_whitelist to check_allowed
|
||||
"""
|
||||
if not self.allowed_users:
|
||||
# No allowed set means any name is allowed
|
||||
return True
|
||||
return username in self.allowed_users
|
||||
|
||||
def check_blocked_users(self, username, authentication=None):
|
||||
"""Check if a username is blocked to authenticate based on Authenticator.blocked configuration
|
||||
|
||||
Return True if username is allowed, False otherwise.
|
||||
No blacklist means any username is allowed.
|
||||
No block list means any username is allowed.
|
||||
|
||||
Names are normalized *before* being checked against the blacklist.
|
||||
Names are normalized *before* being checked against the block list.
|
||||
|
||||
.. versionadded: 0.9
|
||||
|
||||
.. versionchanged:: 1.0
|
||||
Signature updated to accept authentication data as second argument
|
||||
|
||||
.. versionchanged:: 1.2
|
||||
Renamed check_blacklist to check_blocked_users
|
||||
"""
|
||||
if not self.blacklist:
|
||||
# No blacklist means any name is allowed
|
||||
if not self.blocked_users:
|
||||
# No block list means any name is allowed
|
||||
return True
|
||||
return username not in self.blacklist
|
||||
return username not in self.blocked_users
|
||||
|
||||
async def get_authenticated_user(self, handler, data):
|
||||
"""Authenticate the user who is attempting to log in
|
||||
@@ -367,7 +443,7 @@ class Authenticator(LoggingConfigurable):
|
||||
|
||||
This calls `authenticate`, which should be overridden in subclasses,
|
||||
normalizes the username if any normalization should be done,
|
||||
and then validates the name in the whitelist.
|
||||
and then validates the name in the allowed set.
|
||||
|
||||
This is the outer API for authenticating a user.
|
||||
Subclasses should not override this method.
|
||||
@@ -375,7 +451,7 @@ class Authenticator(LoggingConfigurable):
|
||||
The various stages can be overridden separately:
|
||||
- `authenticate` turns formdata into a username
|
||||
- `normalize_username` normalizes the username
|
||||
- `check_whitelist` checks against the user whitelist
|
||||
- `check_allowed` checks against the allowed usernames
|
||||
|
||||
.. versionchanged:: 0.8
|
||||
return dict instead of username
|
||||
@@ -389,7 +465,7 @@ class Authenticator(LoggingConfigurable):
|
||||
else:
|
||||
authenticated = {'name': authenticated}
|
||||
authenticated.setdefault('auth_state', None)
|
||||
# Leave the default as None, but reevaluate later post-whitelist
|
||||
# Leave the default as None, but reevaluate later post-allowed-check
|
||||
authenticated.setdefault('admin', None)
|
||||
|
||||
# normalize the username
|
||||
@@ -400,20 +476,18 @@ class Authenticator(LoggingConfigurable):
|
||||
self.log.warning("Disallowing invalid username %r.", username)
|
||||
return
|
||||
|
||||
blacklist_pass = await maybe_future(
|
||||
self.check_blacklist(username, authenticated)
|
||||
)
|
||||
whitelist_pass = await maybe_future(
|
||||
self.check_whitelist(username, authenticated)
|
||||
blocked_pass = await maybe_future(
|
||||
self.check_blocked_users(username, authenticated)
|
||||
)
|
||||
allowed_pass = await maybe_future(self.check_allowed(username, authenticated))
|
||||
|
||||
if blacklist_pass:
|
||||
if blocked_pass:
|
||||
pass
|
||||
else:
|
||||
self.log.warning("User %r in blacklist. Stop authentication", username)
|
||||
self.log.warning("User %r blocked. Stop authentication", username)
|
||||
return
|
||||
|
||||
if whitelist_pass:
|
||||
if allowed_pass:
|
||||
if authenticated['admin'] is None:
|
||||
authenticated['admin'] = await maybe_future(
|
||||
self.is_admin(handler, authenticated)
|
||||
@@ -423,7 +497,7 @@ class Authenticator(LoggingConfigurable):
|
||||
|
||||
return authenticated
|
||||
else:
|
||||
self.log.warning("User %r not in whitelist.", username)
|
||||
self.log.warning("User %r not allowed.", username)
|
||||
return
|
||||
|
||||
async def refresh_user(self, user, handler=None):
|
||||
@@ -479,7 +553,7 @@ class Authenticator(LoggingConfigurable):
|
||||
It must return the username on successful authentication,
|
||||
and return None on failed authentication.
|
||||
|
||||
Checking the whitelist is handled separately by the caller.
|
||||
Checking allowed_users/blocked_users is handled separately by the caller.
|
||||
|
||||
.. versionchanged:: 0.8
|
||||
Allow `authenticate` to return a dict containing auth_state.
|
||||
@@ -520,10 +594,10 @@ class Authenticator(LoggingConfigurable):
|
||||
|
||||
This method may be a coroutine.
|
||||
|
||||
By default, this just adds the user to the whitelist.
|
||||
By default, this just adds the user to the allowed_users set.
|
||||
|
||||
Subclasses may do more extensive things, such as adding actual unix users,
|
||||
but they should call super to ensure the whitelist is updated.
|
||||
but they should call super to ensure the allowed_users set is updated.
|
||||
|
||||
Note that this should be idempotent, since it is called whenever the hub restarts
|
||||
for all users.
|
||||
@@ -533,19 +607,19 @@ class Authenticator(LoggingConfigurable):
|
||||
"""
|
||||
if not self.validate_username(user.name):
|
||||
raise ValueError("Invalid username: %s" % user.name)
|
||||
if self.whitelist:
|
||||
self.whitelist.add(user.name)
|
||||
if self.allowed_users:
|
||||
self.allowed_users.add(user.name)
|
||||
|
||||
def delete_user(self, user):
|
||||
"""Hook called when a user is deleted
|
||||
|
||||
Removes the user from the whitelist.
|
||||
Subclasses should call super to ensure the whitelist is updated.
|
||||
Removes the user from the allowed_users set.
|
||||
Subclasses should call super to ensure the allowed_users set is updated.
|
||||
|
||||
Args:
|
||||
user (User): The User wrapper object
|
||||
"""
|
||||
self.whitelist.discard(user.name)
|
||||
self.allowed_users.discard(user.name)
|
||||
|
||||
auto_login = Bool(
|
||||
False,
|
||||
@@ -610,6 +684,41 @@ class Authenticator(LoggingConfigurable):
|
||||
return [('/login', LoginHandler)]
|
||||
|
||||
|
||||
def _deprecated_method(old_name, new_name, version):
|
||||
"""Create a deprecated method wrapper for a deprecated method name"""
|
||||
|
||||
def deprecated(self, *args, **kwargs):
|
||||
warnings.warn(
|
||||
(
|
||||
"{cls}.{old_name} is deprecated in JupyterHub {version}."
|
||||
" Please use {cls}.{new_name} instead."
|
||||
).format(
|
||||
cls=self.__class__.__name__,
|
||||
old_name=old_name,
|
||||
new_name=new_name,
|
||||
version=version,
|
||||
),
|
||||
DeprecationWarning,
|
||||
stacklevel=2,
|
||||
)
|
||||
old_method = getattr(self, new_name)
|
||||
return old_method(*args, **kwargs)
|
||||
|
||||
return deprecated
|
||||
|
||||
|
||||
import types
|
||||
|
||||
# deprecate white/blacklist method names
|
||||
for _old_name, _new_name, _version in [
|
||||
("check_whitelist", "check_allowed", "1.2"),
|
||||
("check_blacklist", "check_blocked_users", "1.2"),
|
||||
]:
|
||||
setattr(
|
||||
Authenticator, _old_name, _deprecated_method(_old_name, _new_name, _version),
|
||||
)
|
||||
|
||||
|
||||
class LocalAuthenticator(Authenticator):
|
||||
"""Base class for Authenticators that work with local Linux/UNIX users
|
||||
|
||||
@@ -660,37 +769,46 @@ class LocalAuthenticator(Authenticator):
|
||||
# This appears to be the Linux non-interactive adduser command:
|
||||
return ['adduser', '-q', '--gecos', '""', '--disabled-password']
|
||||
|
||||
group_whitelist = Set(
|
||||
uids = Dict(
|
||||
help="""
|
||||
Whitelist all users from this UNIX group.
|
||||
|
||||
This makes the username whitelist ineffective.
|
||||
Dictionary of uids to use at user creation time.
|
||||
This helps ensure that users created from the database
|
||||
get the same uid each time they are created
|
||||
in temporary deployments or containers.
|
||||
"""
|
||||
).tag(config=True)
|
||||
|
||||
@observe('group_whitelist')
|
||||
def _group_whitelist_changed(self, change):
|
||||
group_whitelist = Set(help="""DEPRECATED: use allowed_groups""",).tag(config=True)
|
||||
|
||||
allowed_groups = Set(
|
||||
help="""
|
||||
Allow login from all users in these UNIX groups.
|
||||
|
||||
If set, allowed username set is ignored.
|
||||
"""
|
||||
Log a warning if both group_whitelist and user whitelist are set.
|
||||
"""
|
||||
if self.whitelist:
|
||||
).tag(config=True)
|
||||
|
||||
@observe('allowed_groups')
|
||||
def _allowed_groups_changed(self, change):
|
||||
"""Log a warning if mutually exclusive user and group allowed sets are specified."""
|
||||
if self.allowed_users:
|
||||
self.log.warning(
|
||||
"Ignoring username whitelist because group whitelist supplied!"
|
||||
"Ignoring Authenticator.allowed_users set because Authenticator.allowed_groups supplied!"
|
||||
)
|
||||
|
||||
def check_whitelist(self, username, authentication=None):
|
||||
if self.group_whitelist:
|
||||
return self.check_group_whitelist(username, authentication)
|
||||
def check_allowed(self, username, authentication=None):
|
||||
if self.allowed_groups:
|
||||
return self.check_allowed_groups(username, authentication)
|
||||
else:
|
||||
return super().check_whitelist(username, authentication)
|
||||
return super().check_allowed(username, authentication)
|
||||
|
||||
def check_group_whitelist(self, username, authentication=None):
|
||||
def check_allowed_groups(self, username, authentication=None):
|
||||
"""
|
||||
If group_whitelist is configured, check if authenticating user is part of group.
|
||||
If allowed_groups is configured, check if authenticating user is part of group.
|
||||
"""
|
||||
if not self.group_whitelist:
|
||||
if not self.allowed_groups:
|
||||
return False
|
||||
for grnam in self.group_whitelist:
|
||||
for grnam in self.allowed_groups:
|
||||
try:
|
||||
group = self._getgrnam(grnam)
|
||||
except KeyError:
|
||||
@@ -762,7 +880,13 @@ class LocalAuthenticator(Authenticator):
|
||||
Tested to work on FreeBSD and Linux, at least.
|
||||
"""
|
||||
name = user.name
|
||||
cmd = [arg.replace('USERNAME', name) for arg in self.add_user_cmd] + [name]
|
||||
cmd = [arg.replace('USERNAME', name) for arg in self.add_user_cmd]
|
||||
try:
|
||||
uid = self.uids[name]
|
||||
cmd += ['--uid', '%d' % uid]
|
||||
except KeyError:
|
||||
self.log.debug("No UID for user %s" % name)
|
||||
cmd += [name]
|
||||
self.log.info("Creating user: %s", ' '.join(map(pipes.quote, cmd)))
|
||||
p = Popen(cmd, stdout=PIPE, stderr=STDOUT)
|
||||
p.wait()
|
||||
@@ -828,7 +952,7 @@ class PAMAuthenticator(LocalAuthenticator):
|
||||
Authoritative list of user groups that determine admin access.
|
||||
Users not in these groups can still be granted admin status through admin_users.
|
||||
|
||||
White/blacklisting rules still apply.
|
||||
allowed/blocked rules still apply.
|
||||
"""
|
||||
).tag(config=True)
|
||||
|
||||
@@ -966,10 +1090,21 @@ class PAMAuthenticator(LocalAuthenticator):
|
||||
uid = pwd.getpwnam(username).pw_uid
|
||||
username = pwd.getpwuid(uid).pw_name
|
||||
username = self.username_map.get(username, username)
|
||||
return username
|
||||
else:
|
||||
return super().normalize_username(username)
|
||||
|
||||
|
||||
for _old_name, _new_name, _version in [
|
||||
("check_group_whitelist", "check_group_allowed", "1.2"),
|
||||
]:
|
||||
setattr(
|
||||
LocalAuthenticator,
|
||||
_old_name,
|
||||
_deprecated_method(_old_name, _new_name, _version),
|
||||
)
|
||||
|
||||
|
||||
class DummyAuthenticator(Authenticator):
|
||||
"""Dummy Authenticator for testing
|
||||
|
||||
|
@@ -6,7 +6,6 @@ from concurrent.futures import ThreadPoolExecutor
|
||||
|
||||
from traitlets import Any
|
||||
from traitlets import default
|
||||
from traitlets import Dict
|
||||
from traitlets import Integer
|
||||
from traitlets import List
|
||||
from traitlets import observe
|
||||
|
59
jupyterhub/event-schemas/server-actions/v1.yaml
Normal file
59
jupyterhub/event-schemas/server-actions/v1.yaml
Normal file
@@ -0,0 +1,59 @@
|
||||
"$id": hub.jupyter.org/server-action
|
||||
version: 1
|
||||
title: JupyterHub server events
|
||||
description: |
|
||||
Record actions on user servers made via JupyterHub.
|
||||
|
||||
JupyterHub can perform various actions on user servers via
|
||||
direct interaction from users, or via the API. This event is
|
||||
recorded whenever either of those happen.
|
||||
|
||||
Limitations:
|
||||
|
||||
1. This does not record all server starts / stops, only those
|
||||
explicitly performed by JupyterHub. For example, a user's server
|
||||
can go down because the node it was running on dies. That will
|
||||
not cause an event to be recorded, since it was not initiated
|
||||
by JupyterHub. In practice this happens often, so this is not
|
||||
a complete record.
|
||||
2. Events are only recorded when an action succeeds.
|
||||
type: object
|
||||
required:
|
||||
- action
|
||||
- username
|
||||
- servername
|
||||
properties:
|
||||
action:
|
||||
enum:
|
||||
- start
|
||||
- stop
|
||||
description: |
|
||||
Action performed by JupyterHub.
|
||||
|
||||
This is a required field.
|
||||
|
||||
Possibl Values:
|
||||
|
||||
1. start
|
||||
A user's server was successfully started
|
||||
|
||||
2. stop
|
||||
A user's server was successfully stopped
|
||||
username:
|
||||
type: string
|
||||
description: |
|
||||
Name of the user whose server this action was performed on.
|
||||
|
||||
This is the normalized name used by JupyterHub itself,
|
||||
which is derived from the authentication provider used but
|
||||
might not be the same as used in the authentication provider.
|
||||
servername:
|
||||
type: string
|
||||
description: |
|
||||
Name of the server this action was performed on.
|
||||
|
||||
JupyterHub supports each user having multiple servers with
|
||||
arbitrary names, and this field specifies the name of the
|
||||
server.
|
||||
|
||||
The 'default' server is denoted by the empty string
|
@@ -2,7 +2,6 @@
|
||||
# Copyright (c) Jupyter Development Team.
|
||||
# Distributed under the terms of the Modified BSD License.
|
||||
import asyncio
|
||||
import copy
|
||||
import json
|
||||
import math
|
||||
import random
|
||||
@@ -26,13 +25,14 @@ from tornado.httputil import HTTPHeaders
|
||||
from tornado.httputil import url_concat
|
||||
from tornado.ioloop import IOLoop
|
||||
from tornado.log import app_log
|
||||
from tornado.web import MissingArgumentError
|
||||
from tornado.web import addslash
|
||||
from tornado.web import RequestHandler
|
||||
|
||||
from .. import __version__
|
||||
from .. import orm
|
||||
from ..metrics import PROXY_ADD_DURATION_SECONDS
|
||||
from ..metrics import ProxyAddStatus
|
||||
from ..metrics import PROXY_DELETE_DURATION_SECONDS
|
||||
from ..metrics import ProxyDeleteStatus
|
||||
from ..metrics import RUNNING_SERVERS
|
||||
from ..metrics import SERVER_POLL_DURATION_SECONDS
|
||||
from ..metrics import SERVER_SPAWN_DURATION_SECONDS
|
||||
@@ -139,6 +139,10 @@ class BaseHandler(RequestHandler):
|
||||
def hub(self):
|
||||
return self.settings['hub']
|
||||
|
||||
@property
|
||||
def app(self):
|
||||
return self.settings['app']
|
||||
|
||||
@property
|
||||
def proxy(self):
|
||||
return self.settings['proxy']
|
||||
@@ -155,6 +159,10 @@ class BaseHandler(RequestHandler):
|
||||
def oauth_provider(self):
|
||||
return self.settings['oauth_provider']
|
||||
|
||||
@property
|
||||
def eventlog(self):
|
||||
return self.settings['eventlog']
|
||||
|
||||
def finish(self, *args, **kwargs):
|
||||
"""Roll back any uncommitted transactions from the handler."""
|
||||
if self.db.dirty:
|
||||
@@ -248,10 +256,40 @@ class BaseHandler(RequestHandler):
|
||||
orm_token = orm.OAuthAccessToken.find(self.db, token)
|
||||
if orm_token is None:
|
||||
return None
|
||||
orm_token.last_activity = orm_token.user.last_activity = datetime.utcnow()
|
||||
self.db.commit()
|
||||
|
||||
now = datetime.utcnow()
|
||||
recorded = self._record_activity(orm_token, now)
|
||||
if self._record_activity(orm_token.user, now) or recorded:
|
||||
self.db.commit()
|
||||
return self._user_from_orm(orm_token.user)
|
||||
|
||||
def _record_activity(self, obj, timestamp=None):
|
||||
"""record activity on an ORM object
|
||||
|
||||
If last_activity was more recent than self.activity_resolution seconds ago,
|
||||
do nothing to avoid unnecessarily frequent database commits.
|
||||
|
||||
Args:
|
||||
obj: an ORM object with a last_activity attribute
|
||||
timestamp (datetime, optional): the timestamp of activity to register.
|
||||
Returns:
|
||||
recorded (bool): True if activity was recorded, False if not.
|
||||
"""
|
||||
if timestamp is None:
|
||||
timestamp = datetime.utcnow()
|
||||
resolution = self.settings.get("activity_resolution", 0)
|
||||
if not obj.last_activity or resolution == 0:
|
||||
self.log.debug("Recording first activity for %s", obj)
|
||||
obj.last_activity = timestamp
|
||||
return True
|
||||
if (timestamp - obj.last_activity).total_seconds() > resolution:
|
||||
# this debug line will happen just too often
|
||||
# uncomment to debug last_activity updates
|
||||
# self.log.debug("Recording activity for %s", obj)
|
||||
obj.last_activity = timestamp
|
||||
return True
|
||||
return False
|
||||
|
||||
async def refresh_auth(self, user, force=False):
|
||||
"""Refresh user authentication info
|
||||
|
||||
@@ -322,14 +360,15 @@ class BaseHandler(RequestHandler):
|
||||
|
||||
# record token activity
|
||||
now = datetime.utcnow()
|
||||
orm_token.last_activity = now
|
||||
recorded = self._record_activity(orm_token, now)
|
||||
if orm_token.user:
|
||||
# FIXME: scopes should give us better control than this
|
||||
# don't consider API requests originating from a server
|
||||
# to be activity from the user
|
||||
if not orm_token.note.startswith("Server at "):
|
||||
orm_token.user.last_activity = now
|
||||
self.db.commit()
|
||||
recorded = self._record_activity(orm_token.user, now) or recorded
|
||||
if recorded:
|
||||
self.db.commit()
|
||||
|
||||
if orm_token.service:
|
||||
return orm_token.service
|
||||
@@ -359,8 +398,8 @@ class BaseHandler(RequestHandler):
|
||||
clear()
|
||||
return
|
||||
# update user activity
|
||||
user.last_activity = datetime.utcnow()
|
||||
self.db.commit()
|
||||
if self._record_activity(user):
|
||||
self.db.commit()
|
||||
return user
|
||||
|
||||
def _user_from_orm(self, orm_user):
|
||||
@@ -452,6 +491,8 @@ class BaseHandler(RequestHandler):
|
||||
path=url_path_join(self.base_url, 'services'),
|
||||
**kwargs
|
||||
)
|
||||
# Reset _jupyterhub_user
|
||||
self._jupyterhub_user = None
|
||||
|
||||
def _set_cookie(self, key, value, encrypted=True, **overrides):
|
||||
"""Setting any cookie should go through here
|
||||
@@ -594,8 +635,15 @@ class BaseHandler(RequestHandler):
|
||||
)
|
||||
|
||||
if not next_url:
|
||||
# custom default URL
|
||||
next_url = default or self.default_url
|
||||
# custom default URL, usually passed because user landed on that page but was not logged in
|
||||
if default:
|
||||
next_url = default
|
||||
else:
|
||||
# As set in jupyterhub_config.py
|
||||
if callable(self.default_url):
|
||||
next_url = self.default_url(self)
|
||||
else:
|
||||
next_url = self.default_url
|
||||
|
||||
if not next_url:
|
||||
# default URL after login
|
||||
@@ -610,8 +658,42 @@ class BaseHandler(RequestHandler):
|
||||
next_url = url_path_join(self.hub.base_url, 'spawn')
|
||||
else:
|
||||
next_url = url_path_join(self.hub.base_url, 'home')
|
||||
|
||||
next_url = self.append_query_parameters(next_url, exclude=['next'])
|
||||
return next_url
|
||||
|
||||
def append_query_parameters(self, url, exclude=None):
|
||||
"""Append the current request's query parameters to the given URL.
|
||||
|
||||
Supports an extra optional parameter ``exclude`` that when provided must
|
||||
contain a list of parameters to be ignored, i.e. these parameters will
|
||||
not be added to the URL.
|
||||
|
||||
This is important to avoid infinite loops with the next parameter being
|
||||
added over and over, for instance.
|
||||
|
||||
The default value for ``exclude`` is an array with "next". This is useful
|
||||
as most use cases in JupyterHub (all?) won't want to include the next
|
||||
parameter twice (the next parameter is added elsewhere to the query
|
||||
parameters).
|
||||
|
||||
:param str url: a URL
|
||||
:param list exclude: optional list of parameters to be ignored, defaults to
|
||||
a list with "next" (to avoid redirect-loops)
|
||||
:rtype (str)
|
||||
"""
|
||||
if exclude is None:
|
||||
exclude = ['next']
|
||||
if self.request.query:
|
||||
query_string = [
|
||||
param
|
||||
for param in parse_qsl(self.request.query)
|
||||
if param[0] not in exclude
|
||||
]
|
||||
if query_string:
|
||||
url = url_concat(url, query_string)
|
||||
return url
|
||||
|
||||
async def auth_to_user(self, authenticated, user=None):
|
||||
"""Persist data from .authenticate() or .refresh_user() to the User database
|
||||
|
||||
@@ -632,9 +714,10 @@ class BaseHandler(RequestHandler):
|
||||
raise ValueError("Username doesn't match! %s != %s" % (username, user.name))
|
||||
|
||||
if user is None:
|
||||
new_user = username not in self.users
|
||||
user = self.user_from_username(username)
|
||||
user = self.find_user(username)
|
||||
new_user = user is None
|
||||
if new_user:
|
||||
user = self.user_from_username(username)
|
||||
await maybe_future(self.authenticator.add_user(user))
|
||||
# Only set `admin` if the authenticator returned an explicit value.
|
||||
if admin is not None and admin != user.admin:
|
||||
@@ -727,6 +810,7 @@ class BaseHandler(RequestHandler):
|
||||
active_counts['spawn_pending'] + active_counts['proxy_pending']
|
||||
)
|
||||
active_count = active_counts['active']
|
||||
RUNNING_SERVERS.set(active_count)
|
||||
|
||||
concurrent_spawn_limit = self.concurrent_spawn_limit
|
||||
active_server_limit = self.active_server_limit
|
||||
@@ -810,10 +894,14 @@ class BaseHandler(RequestHandler):
|
||||
"User %s took %.3f seconds to start", user_server_name, toc - tic
|
||||
)
|
||||
self.statsd.timing('spawner.success', (toc - tic) * 1000)
|
||||
RUNNING_SERVERS.inc()
|
||||
SERVER_SPAWN_DURATION_SECONDS.labels(
|
||||
status=ServerSpawnStatus.success
|
||||
).observe(time.perf_counter() - spawn_start_time)
|
||||
self.eventlog.record_event(
|
||||
'hub.jupyter.org/server-action',
|
||||
1,
|
||||
{'action': 'start', 'username': user.name, 'servername': server_name},
|
||||
)
|
||||
proxy_add_start_time = time.perf_counter()
|
||||
spawner._proxy_pending = True
|
||||
try:
|
||||
@@ -822,12 +910,13 @@ class BaseHandler(RequestHandler):
|
||||
PROXY_ADD_DURATION_SECONDS.labels(status='success').observe(
|
||||
time.perf_counter() - proxy_add_start_time
|
||||
)
|
||||
RUNNING_SERVERS.inc()
|
||||
except Exception:
|
||||
self.log.exception("Failed to add %s to proxy!", user_server_name)
|
||||
self.log.error(
|
||||
"Stopping %s to avoid inconsistent state", user_server_name
|
||||
)
|
||||
await user.stop()
|
||||
await user.stop(server_name)
|
||||
PROXY_ADD_DURATION_SECONDS.labels(status='failure').observe(
|
||||
time.perf_counter() - proxy_add_start_time
|
||||
)
|
||||
@@ -844,7 +933,7 @@ class BaseHandler(RequestHandler):
|
||||
# clear spawner._spawn_future when it's done
|
||||
# keep an exception around, though, to prevent repeated implicit spawns
|
||||
# if spawn is failing
|
||||
if f.exception() is None:
|
||||
if f.cancelled() or f.exception() is None:
|
||||
spawner._spawn_future = None
|
||||
# Now we're all done. clear _spawn_pending flag
|
||||
spawner._spawn_pending = False
|
||||
@@ -855,11 +944,14 @@ class BaseHandler(RequestHandler):
|
||||
# update failure count and abort if consecutive failure limit
|
||||
# is reached
|
||||
def _track_failure_count(f):
|
||||
if f.exception() is None:
|
||||
if f.cancelled() or f.exception() is None:
|
||||
# spawn succeeded, reset failure count
|
||||
self.settings['failure_count'] = 0
|
||||
return
|
||||
# spawn failed, increment count and abort if limit reached
|
||||
SERVER_SPAWN_DURATION_SECONDS.labels(
|
||||
status=ServerSpawnStatus.failure
|
||||
).observe(time.perf_counter() - spawn_start_time)
|
||||
self.settings.setdefault('failure_count', 0)
|
||||
self.settings['failure_count'] += 1
|
||||
failure_count = self.settings['failure_count']
|
||||
@@ -892,13 +984,16 @@ class BaseHandler(RequestHandler):
|
||||
# waiting_for_response indicates server process has started,
|
||||
# but is yet to become responsive.
|
||||
if spawner._spawn_pending and not spawner._waiting_for_response:
|
||||
# still in Spawner.start, which is taking a long time
|
||||
# we shouldn't poll while spawn is incomplete.
|
||||
self.log.warning(
|
||||
"User %s is slow to start (timeout=%s)",
|
||||
user_server_name,
|
||||
self.slow_spawn_timeout,
|
||||
)
|
||||
# If slow_spawn_timeout is intentionally disabled then we
|
||||
# don't need to log a warning, just return.
|
||||
if self.slow_spawn_timeout > 0:
|
||||
# still in Spawner.start, which is taking a long time
|
||||
# we shouldn't poll while spawn is incomplete.
|
||||
self.log.warning(
|
||||
"User %s is slow to start (timeout=%s)",
|
||||
user_server_name,
|
||||
self.slow_spawn_timeout,
|
||||
)
|
||||
return
|
||||
|
||||
# start has finished, but the server hasn't come up
|
||||
@@ -961,7 +1056,18 @@ class BaseHandler(RequestHandler):
|
||||
self.log.warning(
|
||||
"User %s server stopped, with exit code: %s", user.name, status
|
||||
)
|
||||
await self.proxy.delete_user(user, server_name)
|
||||
proxy_deletion_start_time = time.perf_counter()
|
||||
try:
|
||||
await self.proxy.delete_user(user, server_name)
|
||||
PROXY_DELETE_DURATION_SECONDS.labels(
|
||||
status=ProxyDeleteStatus.success
|
||||
).observe(time.perf_counter() - proxy_deletion_start_time)
|
||||
except Exception:
|
||||
PROXY_DELETE_DURATION_SECONDS.labels(
|
||||
status=ProxyDeleteStatus.failure
|
||||
).observe(time.perf_counter() - proxy_deletion_start_time)
|
||||
raise
|
||||
|
||||
await user.stop(server_name)
|
||||
|
||||
async def stop_single_user(self, user, server_name=''):
|
||||
@@ -984,17 +1090,32 @@ class BaseHandler(RequestHandler):
|
||||
tic = time.perf_counter()
|
||||
try:
|
||||
await self.proxy.delete_user(user, server_name)
|
||||
PROXY_DELETE_DURATION_SECONDS.labels(
|
||||
status=ProxyDeleteStatus.success
|
||||
).observe(time.perf_counter() - tic)
|
||||
|
||||
await user.stop(server_name)
|
||||
toc = time.perf_counter()
|
||||
self.log.info(
|
||||
"User %s server took %.3f seconds to stop", user.name, toc - tic
|
||||
)
|
||||
self.statsd.timing('spawner.stop', (toc - tic) * 1000)
|
||||
RUNNING_SERVERS.dec()
|
||||
SERVER_STOP_DURATION_SECONDS.labels(
|
||||
status=ServerStopStatus.success
|
||||
).observe(toc - tic)
|
||||
self.eventlog.record_event(
|
||||
'hub.jupyter.org/server-action',
|
||||
1,
|
||||
{
|
||||
'action': 'stop',
|
||||
'username': user.name,
|
||||
'servername': server_name,
|
||||
},
|
||||
)
|
||||
except:
|
||||
PROXY_DELETE_DURATION_SECONDS.labels(
|
||||
status=ProxyDeleteStatus.failure
|
||||
).observe(time.perf_counter() - tic)
|
||||
SERVER_STOP_DURATION_SECONDS.labels(
|
||||
status=ServerStopStatus.failure
|
||||
).observe(time.perf_counter() - tic)
|
||||
@@ -1009,7 +1130,10 @@ class BaseHandler(RequestHandler):
|
||||
except gen.TimeoutError:
|
||||
# hit timeout, but stop is still pending
|
||||
self.log.warning(
|
||||
"User %s:%s server is slow to stop", user.name, server_name
|
||||
"User %s:%s server is slow to stop (timeout=%s)",
|
||||
user.name,
|
||||
server_name,
|
||||
self.slow_stop_timeout,
|
||||
)
|
||||
|
||||
# return handle on the future for hooking up callbacks
|
||||
@@ -1055,11 +1179,24 @@ class BaseHandler(RequestHandler):
|
||||
logout_url=self.settings['logout_url'],
|
||||
static_url=self.static_url,
|
||||
version_hash=self.version_hash,
|
||||
services=self.get_accessible_services(user),
|
||||
)
|
||||
if self.settings['template_vars']:
|
||||
ns.update(self.settings['template_vars'])
|
||||
return ns
|
||||
|
||||
def get_accessible_services(self, user):
|
||||
accessible_services = []
|
||||
if user is None:
|
||||
return accessible_services
|
||||
for service in self.services.values():
|
||||
if not service.url:
|
||||
continue
|
||||
if not service.display:
|
||||
continue
|
||||
accessible_services.append(service)
|
||||
return accessible_services
|
||||
|
||||
def write_error(self, status_code, **kwargs):
|
||||
"""render custom error pages"""
|
||||
exc_info = kwargs.get('exc_info')
|
||||
@@ -1325,7 +1462,9 @@ class UserUrlHandler(BaseHandler):
|
||||
return
|
||||
|
||||
pending_url = url_concat(
|
||||
url_path_join(self.hub.base_url, 'spawn-pending', user.name, server_name),
|
||||
url_path_join(
|
||||
self.hub.base_url, 'spawn-pending', user.escaped_name, server_name
|
||||
),
|
||||
{'next': self.request.uri},
|
||||
)
|
||||
if spawner.pending or spawner._failed:
|
||||
@@ -1337,13 +1476,20 @@ class UserUrlHandler(BaseHandler):
|
||||
# serve a page prompting for spawn and 503 error
|
||||
# visiting /user/:name no longer triggers implicit spawn
|
||||
# without explicit user action
|
||||
self.set_status(503)
|
||||
spawn_url = url_concat(
|
||||
url_path_join(self.hub.base_url, "spawn", user.name, server_name),
|
||||
url_path_join(self.hub.base_url, "spawn", user.escaped_name, server_name),
|
||||
{"next": self.request.uri},
|
||||
)
|
||||
self.set_status(503)
|
||||
|
||||
auth_state = await user.get_auth_state()
|
||||
html = self.render_template(
|
||||
"not_running.html", user=user, server_name=server_name, spawn_url=spawn_url
|
||||
"not_running.html",
|
||||
user=user,
|
||||
server_name=server_name,
|
||||
spawn_url=spawn_url,
|
||||
auth_state=auth_state,
|
||||
implicit_spawn_seconds=self.settings.get("implicit_spawn_seconds", 0),
|
||||
)
|
||||
self.finish(html)
|
||||
|
||||
@@ -1416,19 +1562,51 @@ class UserRedirectHandler(BaseHandler):
|
||||
|
||||
If the user is not logged in, send to login URL, redirecting back here.
|
||||
|
||||
If c.JupyterHub.user_redirect_hook is set, the return value of that
|
||||
callable is used to generate the redirect URL.
|
||||
|
||||
.. versionadded:: 0.7
|
||||
"""
|
||||
|
||||
@web.authenticated
|
||||
def get(self, path):
|
||||
user = self.current_user
|
||||
user_url = url_path_join(user.url, path)
|
||||
if self.request.query:
|
||||
user_url = url_concat(user_url, parse_qsl(self.request.query))
|
||||
async def get(self, path):
|
||||
# If hook is present to generate URL to redirect to, use that instead
|
||||
# of the default. The configurer is responsible for making sure this
|
||||
# URL is right. If None is returned by the hook, we do our normal
|
||||
# processing
|
||||
url = None
|
||||
if self.app.user_redirect_hook:
|
||||
url = await maybe_future(
|
||||
self.app.user_redirect_hook(
|
||||
path, self.request, self.current_user, self.base_url
|
||||
)
|
||||
)
|
||||
if url is None:
|
||||
user = self.current_user
|
||||
user_url = user.url
|
||||
|
||||
url = url_concat(
|
||||
url_path_join(self.hub.base_url, "spawn", user.name), {"next": user_url}
|
||||
)
|
||||
if self.app.default_server_name:
|
||||
user_url = url_path_join(user_url, self.app.default_server_name)
|
||||
|
||||
user_url = url_path_join(user_url, path)
|
||||
if self.request.query:
|
||||
user_url = url_concat(user_url, parse_qsl(self.request.query))
|
||||
|
||||
if self.app.default_server_name:
|
||||
url = url_concat(
|
||||
url_path_join(
|
||||
self.hub.base_url,
|
||||
"spawn",
|
||||
user.escaped_name,
|
||||
self.app.default_server_name,
|
||||
),
|
||||
{"next": user_url},
|
||||
)
|
||||
else:
|
||||
url = url_concat(
|
||||
url_path_join(self.hub.base_url, "spawn", user.escaped_name),
|
||||
{"next": user_url},
|
||||
)
|
||||
|
||||
self.redirect(url)
|
||||
|
||||
@@ -1450,10 +1628,9 @@ class CSPReportHandler(BaseHandler):
|
||||
class AddSlashHandler(BaseHandler):
|
||||
"""Handler for adding trailing slash to URLs that need them"""
|
||||
|
||||
def get(self, *args):
|
||||
src = urlparse(self.request.uri)
|
||||
dest = src._replace(path=src.path + '/')
|
||||
self.redirect(urlunparse(dest))
|
||||
@addslash
|
||||
def get(self):
|
||||
pass
|
||||
|
||||
|
||||
default_handlers = [
|
||||
|
@@ -18,33 +18,73 @@ class LogoutHandler(BaseHandler):
|
||||
def shutdown_on_logout(self):
|
||||
return self.settings.get('shutdown_on_logout', False)
|
||||
|
||||
async def get(self):
|
||||
async def _shutdown_servers(self, user):
|
||||
"""Shutdown servers for logout
|
||||
|
||||
Get all active servers for the provided user, stop them.
|
||||
"""
|
||||
active_servers = [
|
||||
name
|
||||
for (name, spawner) in user.spawners.items()
|
||||
if spawner.active and not spawner.pending
|
||||
]
|
||||
if active_servers:
|
||||
self.log.info("Shutting down %s's servers", user.name)
|
||||
futures = []
|
||||
for server_name in active_servers:
|
||||
futures.append(maybe_future(self.stop_single_user(user, server_name)))
|
||||
await asyncio.gather(*futures)
|
||||
|
||||
def _backend_logout_cleanup(self, name):
|
||||
"""Default backend logout actions
|
||||
|
||||
Send a log message, clear some cookies, increment the logout counter.
|
||||
"""
|
||||
self.log.info("User logged out: %s", name)
|
||||
self.clear_login_cookie()
|
||||
self.statsd.incr('logout')
|
||||
|
||||
async def default_handle_logout(self):
|
||||
"""The default logout action
|
||||
|
||||
Optionally cleans up servers, clears cookies, increments logout counter
|
||||
Cleaning up servers can be prevented by setting shutdown_on_logout to
|
||||
False.
|
||||
"""
|
||||
user = self.current_user
|
||||
if user:
|
||||
if self.shutdown_on_logout:
|
||||
active_servers = [
|
||||
name
|
||||
for (name, spawner) in user.spawners.items()
|
||||
if spawner.active and not spawner.pending
|
||||
]
|
||||
if active_servers:
|
||||
self.log.info("Shutting down %s's servers", user.name)
|
||||
futures = []
|
||||
for server_name in active_servers:
|
||||
futures.append(
|
||||
maybe_future(self.stop_single_user(user, server_name))
|
||||
)
|
||||
await asyncio.gather(*futures)
|
||||
await self._shutdown_servers(user)
|
||||
|
||||
self.log.info("User logged out: %s", user.name)
|
||||
self.clear_login_cookie()
|
||||
self.statsd.incr('logout')
|
||||
self._backend_logout_cleanup(user.name)
|
||||
|
||||
async def handle_logout(self):
|
||||
"""Custom user action during logout
|
||||
|
||||
By default a no-op, this function should be overridden in subclasses
|
||||
to have JupyterHub take a custom action on logout.
|
||||
"""
|
||||
return
|
||||
|
||||
async def render_logout_page(self):
|
||||
"""Render the logout page, if any
|
||||
|
||||
Override this function to set a custom logout page.
|
||||
"""
|
||||
if self.authenticator.auto_login:
|
||||
html = self.render_template('logout.html')
|
||||
self.finish(html)
|
||||
else:
|
||||
self.redirect(self.settings['login_url'], permanent=False)
|
||||
|
||||
async def get(self):
|
||||
"""Log the user out, call the custom action, forward the user
|
||||
to the logout page
|
||||
"""
|
||||
await self.default_handle_logout()
|
||||
await self.handle_logout()
|
||||
await self.render_logout_page()
|
||||
|
||||
|
||||
class LoginHandler(BaseHandler):
|
||||
"""Render the login page."""
|
||||
|
@@ -1,7 +1,6 @@
|
||||
from prometheus_client import CONTENT_TYPE_LATEST
|
||||
from prometheus_client import generate_latest
|
||||
from prometheus_client import REGISTRY
|
||||
from tornado import gen
|
||||
|
||||
from ..utils import metrics_authentication
|
||||
from .base import BaseHandler
|
||||
|
@@ -2,20 +2,21 @@
|
||||
# Copyright (c) Jupyter Development Team.
|
||||
# Distributed under the terms of the Modified BSD License.
|
||||
import asyncio
|
||||
import copy
|
||||
import time
|
||||
from collections import defaultdict
|
||||
from datetime import datetime
|
||||
from http.client import responses
|
||||
|
||||
from jinja2 import TemplateNotFound
|
||||
from tornado import gen
|
||||
from tornado import web
|
||||
from tornado.httputil import url_concat
|
||||
from tornado.httputil import urlparse
|
||||
|
||||
from .. import __version__
|
||||
from .. import orm
|
||||
from ..metrics import SERVER_POLL_DURATION_SECONDS
|
||||
from ..metrics import ServerPollStatus
|
||||
from ..pagination import Pagination
|
||||
from ..utils import admin_only
|
||||
from ..utils import maybe_future
|
||||
from ..utils import url_path_join
|
||||
@@ -61,12 +62,14 @@ class HomeHandler(BaseHandler):
|
||||
# to establish that this is an explicit spawn request rather
|
||||
# than an implicit one, which can be caused by any link to `/user/:name(/:server_name)`
|
||||
if user.active:
|
||||
url = url_path_join(self.base_url, 'user', user.name)
|
||||
url = url_path_join(self.base_url, 'user', user.escaped_name)
|
||||
else:
|
||||
url = url_path_join(self.hub.base_url, 'spawn', user.name)
|
||||
url = url_path_join(self.hub.base_url, 'spawn', user.escaped_name)
|
||||
|
||||
auth_state = await user.get_auth_state()
|
||||
html = self.render_template(
|
||||
'home.html',
|
||||
auth_state=auth_state,
|
||||
user=user,
|
||||
url=url,
|
||||
allow_named_servers=self.allow_named_servers,
|
||||
@@ -89,10 +92,12 @@ class SpawnHandler(BaseHandler):
|
||||
|
||||
default_url = None
|
||||
|
||||
def _render_form(self, for_user, spawner_options_form, message=''):
|
||||
async def _render_form(self, for_user, spawner_options_form, message=''):
|
||||
auth_state = await for_user.get_auth_state()
|
||||
return self.render_template(
|
||||
'spawn.html',
|
||||
for_user=for_user,
|
||||
auth_state=auth_state,
|
||||
spawner_options_form=spawner_options_form,
|
||||
error_message=message,
|
||||
url=self.request.uri,
|
||||
@@ -117,6 +122,23 @@ class SpawnHandler(BaseHandler):
|
||||
if user is None:
|
||||
raise web.HTTPError(404, "No such user: %s" % for_user)
|
||||
|
||||
if server_name:
|
||||
if not self.allow_named_servers:
|
||||
raise web.HTTPError(400, "Named servers are not enabled.")
|
||||
if (
|
||||
self.named_server_limit_per_user > 0
|
||||
and server_name not in user.orm_spawners
|
||||
):
|
||||
named_spawners = list(user.all_spawners(include_default=False))
|
||||
if self.named_server_limit_per_user <= len(named_spawners):
|
||||
raise web.HTTPError(
|
||||
400,
|
||||
"User {} already has the maximum of {} named servers."
|
||||
" One must be deleted before a new server can be created".format(
|
||||
user.name, self.named_server_limit_per_user
|
||||
),
|
||||
)
|
||||
|
||||
if not self.allow_named_servers and user.running:
|
||||
url = self.get_next_url(user, default=user.server_url(server_name))
|
||||
self.log.info("User is running: %s", user.name)
|
||||
@@ -127,17 +149,8 @@ class SpawnHandler(BaseHandler):
|
||||
server_name = ''
|
||||
|
||||
spawner = user.spawners[server_name]
|
||||
# resolve `?next=...`, falling back on the spawn-pending url
|
||||
# must not be /user/server for named servers,
|
||||
# which may get handled by the default server if they aren't ready yet
|
||||
|
||||
pending_url = url_path_join(
|
||||
self.hub.base_url, "spawn-pending", user.name, server_name
|
||||
)
|
||||
|
||||
if self.get_argument('next', None):
|
||||
# preserve `?next=...` through spawn-pending
|
||||
pending_url = url_concat(pending_url, {'next': self.get_argument('next')})
|
||||
pending_url = self._get_pending_url(user, server_name)
|
||||
|
||||
# spawner is active, redirect back to get progress, etc.
|
||||
if spawner.ready:
|
||||
@@ -153,26 +166,56 @@ class SpawnHandler(BaseHandler):
|
||||
|
||||
# Add handler to spawner here so you can access query params in form rendering.
|
||||
spawner.handler = self
|
||||
|
||||
# auth_state may be an input to options form,
|
||||
# so resolve the auth state hook here
|
||||
auth_state = await user.get_auth_state()
|
||||
await spawner.run_auth_state_hook(auth_state)
|
||||
|
||||
# Try to start server directly when query arguments are passed.
|
||||
error_message = ''
|
||||
query_options = {}
|
||||
for key, byte_list in self.request.query_arguments.items():
|
||||
query_options[key] = [bs.decode('utf8') for bs in byte_list]
|
||||
|
||||
# 'next' is reserved argument for redirect after spawn
|
||||
query_options.pop('next', None)
|
||||
|
||||
if len(query_options) > 0:
|
||||
try:
|
||||
self.log.debug(
|
||||
"Triggering spawn with supplied query arguments for %s",
|
||||
spawner._log_name,
|
||||
)
|
||||
options = await maybe_future(spawner.options_from_query(query_options))
|
||||
pending_url = self._get_pending_url(user, server_name)
|
||||
return await self._wrap_spawn_single_user(
|
||||
user, server_name, spawner, pending_url, options
|
||||
)
|
||||
except Exception as e:
|
||||
self.log.error(
|
||||
"Failed to spawn single-user server with query arguments",
|
||||
exc_info=True,
|
||||
)
|
||||
error_message = str(e)
|
||||
# fallback to behavior without failing query arguments
|
||||
|
||||
spawner_options_form = await spawner.get_options_form()
|
||||
if spawner_options_form:
|
||||
self.log.debug("Serving options form for %s", spawner._log_name)
|
||||
form = self._render_form(
|
||||
for_user=user, spawner_options_form=spawner_options_form
|
||||
form = await self._render_form(
|
||||
for_user=user,
|
||||
spawner_options_form=spawner_options_form,
|
||||
message=error_message,
|
||||
)
|
||||
self.finish(form)
|
||||
else:
|
||||
self.log.debug(
|
||||
"Triggering spawn with default options for %s", spawner._log_name
|
||||
)
|
||||
# Explicit spawn request: clear _spawn_future
|
||||
# which may have been saved to prevent implicit spawns
|
||||
# after a failure.
|
||||
if spawner._spawn_future and spawner._spawn_future.done():
|
||||
spawner._spawn_future = None
|
||||
# not running, no form. Trigger spawn and redirect back to /user/:name
|
||||
f = asyncio.ensure_future(self.spawn_single_user(user, server_name))
|
||||
await asyncio.wait([f], timeout=1)
|
||||
self.redirect(pending_url)
|
||||
return await self._wrap_spawn_single_user(
|
||||
user, server_name, spawner, pending_url
|
||||
)
|
||||
|
||||
@web.authenticated
|
||||
async def post(self, for_user=None, server_name=''):
|
||||
@@ -202,14 +245,20 @@ class SpawnHandler(BaseHandler):
|
||||
for key, byte_list in self.request.files.items():
|
||||
form_options["%s_file" % key] = byte_list
|
||||
try:
|
||||
self.log.debug(
|
||||
"Triggering spawn with supplied form options for %s", spawner._log_name
|
||||
)
|
||||
options = await maybe_future(spawner.options_from_form(form_options))
|
||||
await self.spawn_single_user(user, server_name=server_name, options=options)
|
||||
pending_url = self._get_pending_url(user, server_name)
|
||||
return await self._wrap_spawn_single_user(
|
||||
user, server_name, spawner, pending_url, options
|
||||
)
|
||||
except Exception as e:
|
||||
self.log.error(
|
||||
"Failed to spawn single-user server with form", exc_info=True
|
||||
)
|
||||
spawner_options_form = await user.spawner.get_options_form()
|
||||
form = self._render_form(
|
||||
form = await self._render_form(
|
||||
for_user=user, spawner_options_form=spawner_options_form, message=str(e)
|
||||
)
|
||||
self.finish(form)
|
||||
@@ -219,11 +268,52 @@ class SpawnHandler(BaseHandler):
|
||||
next_url = self.get_next_url(
|
||||
user,
|
||||
default=url_path_join(
|
||||
self.hub.base_url, "spawn-pending", user.name, server_name
|
||||
self.hub.base_url, "spawn-pending", user.escaped_name, server_name
|
||||
),
|
||||
)
|
||||
self.redirect(next_url)
|
||||
|
||||
def _get_pending_url(self, user, server_name):
|
||||
# resolve `?next=...`, falling back on the spawn-pending url
|
||||
# must not be /user/server for named servers,
|
||||
# which may get handled by the default server if they aren't ready yet
|
||||
|
||||
pending_url = url_path_join(
|
||||
self.hub.base_url, "spawn-pending", user.escaped_name, server_name
|
||||
)
|
||||
|
||||
pending_url = self.append_query_parameters(pending_url, exclude=['next'])
|
||||
|
||||
if self.get_argument('next', None):
|
||||
# preserve `?next=...` through spawn-pending
|
||||
pending_url = url_concat(pending_url, {'next': self.get_argument('next')})
|
||||
|
||||
return pending_url
|
||||
|
||||
async def _wrap_spawn_single_user(
|
||||
self, user, server_name, spawner, pending_url, options=None
|
||||
):
|
||||
# Explicit spawn request: clear _spawn_future
|
||||
# which may have been saved to prevent implicit spawns
|
||||
# after a failure.
|
||||
if spawner._spawn_future and spawner._spawn_future.done():
|
||||
spawner._spawn_future = None
|
||||
# not running, no form. Trigger spawn and redirect back to /user/:name
|
||||
f = asyncio.ensure_future(
|
||||
self.spawn_single_user(user, server_name, options=options)
|
||||
)
|
||||
done, pending = await asyncio.wait([f], timeout=1)
|
||||
# If spawn_single_user throws an exception, raise a 500 error
|
||||
# otherwise it may cause a redirect loop
|
||||
if f.done() and f.exception():
|
||||
exc = f.exception()
|
||||
raise web.HTTPError(
|
||||
500,
|
||||
"Error in Authenticator.pre_spawn_start: %s %s"
|
||||
% (type(exc).__name__, str(exc)),
|
||||
)
|
||||
return self.redirect(pending_url)
|
||||
|
||||
|
||||
class SpawnPendingHandler(BaseHandler):
|
||||
"""Handle /hub/spawn-pending/:user/:server
|
||||
@@ -270,6 +360,8 @@ class SpawnPendingHandler(BaseHandler):
|
||||
# if spawning fails for any reason, point users to /hub/home to retry
|
||||
self.extra_error_html = self.spawn_home_error
|
||||
|
||||
auth_state = await user.get_auth_state()
|
||||
|
||||
# First, check for previous failure.
|
||||
if (
|
||||
not spawner.active
|
||||
@@ -282,11 +374,14 @@ class SpawnPendingHandler(BaseHandler):
|
||||
# We should point the user to Home if the most recent spawn failed.
|
||||
exc = spawner._spawn_future.exception()
|
||||
self.log.error("Previous spawn for %s failed: %s", spawner._log_name, exc)
|
||||
spawn_url = url_path_join(self.hub.base_url, "spawn", user.escaped_name)
|
||||
spawn_url = url_path_join(
|
||||
self.hub.base_url, "spawn", user.escaped_name, server_name
|
||||
)
|
||||
self.set_status(500)
|
||||
html = self.render_template(
|
||||
"not_running.html",
|
||||
user=user,
|
||||
auth_state=auth_state,
|
||||
server_name=server_name,
|
||||
spawn_url=spawn_url,
|
||||
failed=True,
|
||||
@@ -308,7 +403,11 @@ class SpawnPendingHandler(BaseHandler):
|
||||
else:
|
||||
page = "spawn_pending.html"
|
||||
html = self.render_template(
|
||||
page, user=user, spawner=spawner, progress_url=spawner._progress_url
|
||||
page,
|
||||
user=user,
|
||||
spawner=spawner,
|
||||
progress_url=spawner._progress_url,
|
||||
auth_state=auth_state,
|
||||
)
|
||||
self.finish(html)
|
||||
return
|
||||
@@ -327,10 +426,13 @@ class SpawnPendingHandler(BaseHandler):
|
||||
# further, set status to 404 because this is not
|
||||
# serving the expected page
|
||||
if status is not None:
|
||||
spawn_url = url_path_join(self.hub.base_url, "spawn", user.escaped_name)
|
||||
spawn_url = url_path_join(
|
||||
self.hub.base_url, "spawn", user.escaped_name, server_name
|
||||
)
|
||||
html = self.render_template(
|
||||
"not_running.html",
|
||||
user=user,
|
||||
auth_state=auth_state,
|
||||
server_name=server_name,
|
||||
spawn_url=spawn_url,
|
||||
)
|
||||
@@ -348,14 +450,18 @@ class SpawnPendingHandler(BaseHandler):
|
||||
class AdminHandler(BaseHandler):
|
||||
"""Render the admin page."""
|
||||
|
||||
@web.authenticated
|
||||
@admin_only
|
||||
def get(self):
|
||||
async def get(self):
|
||||
page, per_page, offset = Pagination(config=self.config).get_page_args(self)
|
||||
|
||||
available = {'name', 'admin', 'running', 'last_activity'}
|
||||
default_sort = ['admin', 'name']
|
||||
mapping = {'running': orm.Spawner.server_id}
|
||||
for name in available:
|
||||
if name not in mapping:
|
||||
mapping[name] = getattr(orm.User, name)
|
||||
table = orm.User if name != "last_activity" else orm.Spawner
|
||||
mapping[name] = getattr(table, name)
|
||||
|
||||
default_order = {
|
||||
'name': 'asc',
|
||||
@@ -390,23 +496,41 @@ class AdminHandler(BaseHandler):
|
||||
# get User.col.desc() order objects
|
||||
ordered = [getattr(c, o)() for c, o in zip(cols, orders)]
|
||||
|
||||
users = self.db.query(orm.User).outerjoin(orm.Spawner).order_by(*ordered)
|
||||
users = (
|
||||
self.db.query(orm.User)
|
||||
.outerjoin(orm.Spawner)
|
||||
.order_by(*ordered)
|
||||
.limit(per_page)
|
||||
.offset(offset)
|
||||
)
|
||||
users = [self._user_from_orm(u) for u in users]
|
||||
from itertools import chain
|
||||
|
||||
running = []
|
||||
for u in users:
|
||||
running.extend(s for s in u.spawners.values() if s.active)
|
||||
|
||||
total = self.db.query(orm.User.id).count()
|
||||
pagination = Pagination(
|
||||
url=self.request.uri,
|
||||
total=total,
|
||||
page=page,
|
||||
per_page=per_page,
|
||||
config=self.config,
|
||||
)
|
||||
|
||||
auth_state = await self.current_user.get_auth_state()
|
||||
html = self.render_template(
|
||||
'admin.html',
|
||||
current_user=self.current_user,
|
||||
auth_state=auth_state,
|
||||
admin_access=self.settings.get('admin_access', False),
|
||||
users=users,
|
||||
running=running,
|
||||
sort={s: o for s, o in zip(sorts, orders)},
|
||||
allow_named_servers=self.allow_named_servers,
|
||||
named_server_limit_per_user=self.named_server_limit_per_user,
|
||||
server_version='{} {}'.format(__version__, self.version_hash),
|
||||
pagination=pagination,
|
||||
)
|
||||
self.finish(html)
|
||||
|
||||
@@ -415,7 +539,7 @@ class TokenPageHandler(BaseHandler):
|
||||
"""Handler for page requesting new API tokens"""
|
||||
|
||||
@web.authenticated
|
||||
def get(self):
|
||||
async def get(self):
|
||||
never = datetime(1900, 1, 1)
|
||||
|
||||
user = self.current_user
|
||||
@@ -484,8 +608,12 @@ class TokenPageHandler(BaseHandler):
|
||||
|
||||
oauth_clients = sorted(oauth_clients, key=sort_key, reverse=True)
|
||||
|
||||
auth_state = await self.current_user.get_auth_state()
|
||||
html = self.render_template(
|
||||
'token.html', api_tokens=api_tokens, oauth_clients=oauth_clients
|
||||
'token.html',
|
||||
api_tokens=api_tokens,
|
||||
oauth_clients=oauth_clients,
|
||||
auth_state=auth_state,
|
||||
)
|
||||
self.finish(html)
|
||||
|
||||
@@ -526,10 +654,14 @@ class ProxyErrorHandler(BaseHandler):
|
||||
|
||||
|
||||
class HealthCheckHandler(BaseHandler):
|
||||
"""Answer to health check"""
|
||||
"""Serve health check probes as quickly as possible"""
|
||||
|
||||
def get(self, *args):
|
||||
self.finish()
|
||||
# There is nothing for us to do other than return a positive
|
||||
# HTTP status code as quickly as possible for GET or HEAD requests
|
||||
def get(self):
|
||||
pass
|
||||
|
||||
head = get
|
||||
|
||||
|
||||
default_handlers = [
|
||||
|
@@ -12,6 +12,7 @@ from tornado.log import LogFormatter
|
||||
from tornado.web import HTTPError
|
||||
from tornado.web import StaticFileHandler
|
||||
|
||||
from .handlers.pages import HealthCheckHandler
|
||||
from .metrics import prometheus_log_method
|
||||
|
||||
|
||||
@@ -98,8 +99,12 @@ def _scrub_headers(headers):
|
||||
headers = dict(headers)
|
||||
if 'Authorization' in headers:
|
||||
auth = headers['Authorization']
|
||||
if auth.startswith('token '):
|
||||
headers['Authorization'] = 'token [secret]'
|
||||
if ' ' in auth:
|
||||
auth_type = auth.split(' ', 1)[0]
|
||||
else:
|
||||
# no space, hide the whole thing in case there was a mistake
|
||||
auth_type = ''
|
||||
headers['Authorization'] = '{} [secret]'.format(auth_type)
|
||||
if 'Cookie' in headers:
|
||||
c = SimpleCookie(headers['Cookie'])
|
||||
redacted = []
|
||||
@@ -123,7 +128,9 @@ def log_request(handler):
|
||||
"""
|
||||
status = handler.get_status()
|
||||
request = handler.request
|
||||
if status == 304 or (status < 300 and isinstance(handler, StaticFileHandler)):
|
||||
if status == 304 or (
|
||||
status < 300 and isinstance(handler, (StaticFileHandler, HealthCheckHandler))
|
||||
):
|
||||
# static-file success and 304 Found are debug-level
|
||||
log_method = access_log.debug
|
||||
elif status < 400:
|
||||
@@ -162,7 +169,7 @@ def log_request(handler):
|
||||
location='',
|
||||
)
|
||||
msg = "{status} {method} {uri}{location} ({user}@{ip}) {request_time:.2f}ms"
|
||||
if status >= 500 and status != 502:
|
||||
if status >= 500 and status not in {502, 503}:
|
||||
log_method(json.dumps(headers, indent=2))
|
||||
elif status in {301, 302}:
|
||||
# log redirect targets
|
||||
|
@@ -39,16 +39,24 @@ RUNNING_SERVERS = Gauge(
|
||||
'running_servers', 'the number of user servers currently running'
|
||||
)
|
||||
|
||||
RUNNING_SERVERS.set(0)
|
||||
|
||||
TOTAL_USERS = Gauge('total_users', 'toal number of users')
|
||||
|
||||
TOTAL_USERS.set(0)
|
||||
TOTAL_USERS = Gauge('total_users', 'total number of users')
|
||||
|
||||
CHECK_ROUTES_DURATION_SECONDS = Histogram(
|
||||
'check_routes_duration_seconds', 'Time taken to validate all routes in proxy'
|
||||
)
|
||||
|
||||
HUB_STARTUP_DURATION_SECONDS = Histogram(
|
||||
'hub_startup_duration_seconds', 'Time taken for Hub to start'
|
||||
)
|
||||
|
||||
INIT_SPAWNERS_DURATION_SECONDS = Histogram(
|
||||
'init_spawners_duration_seconds', 'Time taken for spawners to initialize'
|
||||
)
|
||||
|
||||
PROXY_POLL_DURATION_SECONDS = Histogram(
|
||||
'proxy_poll_duration_seconds', 'duration for polling all routes from proxy'
|
||||
)
|
||||
|
||||
|
||||
class ServerSpawnStatus(Enum):
|
||||
"""
|
||||
@@ -139,14 +147,37 @@ for s in ServerStopStatus:
|
||||
SERVER_STOP_DURATION_SECONDS.labels(status=s)
|
||||
|
||||
|
||||
PROXY_DELETE_DURATION_SECONDS = Histogram(
|
||||
'proxy_delete_duration_seconds',
|
||||
'duration for deleting user routes from proxy',
|
||||
['status'],
|
||||
)
|
||||
|
||||
|
||||
class ProxyDeleteStatus(Enum):
|
||||
"""
|
||||
Possible values for 'status' label of PROXY_DELETE_DURATION_SECONDS
|
||||
"""
|
||||
|
||||
success = 'success'
|
||||
failure = 'failure'
|
||||
|
||||
def __str__(self):
|
||||
return self.value
|
||||
|
||||
|
||||
for s in ProxyDeleteStatus:
|
||||
PROXY_DELETE_DURATION_SECONDS.labels(status=s)
|
||||
|
||||
|
||||
def prometheus_log_method(handler):
|
||||
"""
|
||||
Tornado log handler for recording RED metrics.
|
||||
|
||||
We record the following metrics:
|
||||
Rate – the number of requests, per second, your services are serving.
|
||||
Errors – the number of failed requests per second.
|
||||
Duration – The amount of time each request takes expressed as a time interval.
|
||||
Rate: the number of requests, per second, your services are serving.
|
||||
Errors: the number of failed requests per second.
|
||||
Duration: the amount of time each request takes expressed as a time interval.
|
||||
|
||||
We use a fully qualified name of the handler as a label,
|
||||
rather than every url path to reduce cardinality.
|
||||
|
@@ -2,16 +2,11 @@
|
||||
|
||||
implements https://oauthlib.readthedocs.io/en/latest/oauth2/server.html
|
||||
"""
|
||||
from datetime import datetime
|
||||
from urllib.parse import urlparse
|
||||
|
||||
from oauthlib import uri_validate
|
||||
from oauthlib.oauth2 import RequestValidator
|
||||
from oauthlib.oauth2 import WebApplicationServer
|
||||
from oauthlib.oauth2.rfc6749.grant_types import authorization_code
|
||||
from oauthlib.oauth2.rfc6749.grant_types import base
|
||||
from sqlalchemy.orm import scoped_session
|
||||
from tornado import web
|
||||
from tornado.escape import url_escape
|
||||
from tornado.log import app_log
|
||||
|
||||
@@ -250,7 +245,7 @@ class JupyterHubRequestValidator(RequestValidator):
|
||||
client=orm_client,
|
||||
code=code['code'],
|
||||
# oauth has 5 minutes to complete
|
||||
expires_at=int(datetime.utcnow().timestamp() + 300),
|
||||
expires_at=int(orm.OAuthCode.now() + 300),
|
||||
# TODO: persist oauth scopes
|
||||
# scopes=request.scopes,
|
||||
user=request.user.orm_user,
|
||||
@@ -347,7 +342,7 @@ class JupyterHubRequestValidator(RequestValidator):
|
||||
orm_access_token = orm.OAuthAccessToken(
|
||||
client=client,
|
||||
grant_type=orm.GrantType.authorization_code,
|
||||
expires_at=datetime.utcnow().timestamp() + token['expires_in'],
|
||||
expires_at=orm.OAuthAccessToken.now() + token['expires_in'],
|
||||
refresh_token=token['refresh_token'],
|
||||
# TODO: save scopes,
|
||||
# scopes=scopes,
|
||||
@@ -441,7 +436,7 @@ class JupyterHubRequestValidator(RequestValidator):
|
||||
Method is used by:
|
||||
- Authorization Code Grant
|
||||
"""
|
||||
orm_code = self.db.query(orm.OAuthCode).filter_by(code=code).first()
|
||||
orm_code = orm.OAuthCode.find(self.db, code=code)
|
||||
if orm_code is None:
|
||||
app_log.debug("No such code: %s", code)
|
||||
return False
|
||||
|
@@ -53,7 +53,7 @@ class Server(HasTraits):
|
||||
Never used in APIs, only logging,
|
||||
since it can be non-connectable value, such as '', meaning all interfaces.
|
||||
"""
|
||||
if self.ip in {'', '0.0.0.0'}:
|
||||
if self.ip in {'', '0.0.0.0', '::'}:
|
||||
return self.url.replace(self._connect_ip, self.ip or '*', 1)
|
||||
return self.url
|
||||
|
||||
@@ -87,13 +87,13 @@ class Server(HasTraits):
|
||||
"""The address to use when connecting to this server
|
||||
|
||||
When `ip` is set to a real ip address, the same value is used.
|
||||
When `ip` refers to 'all interfaces' (e.g. '0.0.0.0'),
|
||||
When `ip` refers to 'all interfaces' (e.g. '0.0.0.0' or '::'),
|
||||
clients connect via hostname by default.
|
||||
Setting `connect_ip` explicitly overrides any default behavior.
|
||||
"""
|
||||
if self.connect_ip:
|
||||
return self.connect_ip
|
||||
elif self.ip in {'', '0.0.0.0'}:
|
||||
elif self.ip in {'', '0.0.0.0', '::'}:
|
||||
# if listening on all interfaces, default to hostname for connect
|
||||
return socket.gethostname()
|
||||
else:
|
||||
@@ -149,7 +149,12 @@ class Server(HasTraits):
|
||||
if self.connect_url:
|
||||
parsed = urlparse(self.connect_url)
|
||||
return "{proto}://{host}".format(proto=parsed.scheme, host=parsed.netloc)
|
||||
return "{proto}://{ip}:{port}".format(
|
||||
|
||||
if ':' in self._connect_ip:
|
||||
fmt = "{proto}://[{ip}]:{port}"
|
||||
else:
|
||||
fmt = "{proto}://{ip}:{port}"
|
||||
return fmt.format(
|
||||
proto=self.proto, ip=self._connect_ip, port=self._connect_port
|
||||
)
|
||||
|
||||
@@ -213,8 +218,4 @@ class Hub(Server):
|
||||
return url_path_join(self.url, 'api')
|
||||
|
||||
def __repr__(self):
|
||||
return "<%s %s:%s>" % (
|
||||
self.__class__.__name__,
|
||||
self.server.ip,
|
||||
self.server.port,
|
||||
)
|
||||
return "<%s %s:%s>" % (self.__class__.__name__, self.ip, self.port)
|
||||
|
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user