You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
|Marius van Niekerk | MvN | mariusvniekerk | Voltron Data/CF|
19
+
|Cheng H. Lee| CHL| chenghlee| Anaconda/CF|
20
20
|||||
21
21
|||||
22
22
|||||
23
23
|||||
24
24
|||||
25
25
26
-
X people total
26
+
12 people total
27
27
28
28
### Standing items
29
29
@@ -37,9 +37,28 @@ X people total
37
37
38
38
-[ ]
39
39
40
-
### Your __new__() agenda items
41
-
42
-
-[ ]
40
+
### Your `__new__()` agenda items
41
+
42
+
-[x] (HV) How do we introduce `{{ stdlib("c") }}`?
43
+
- Design [objections](https://github.com/conda/conda-build/issues/5053) were withdrawn, functionality is released in conda-build 3.28 already.
44
+
- Main question is how to roll out this change broadly. I suggest to keep it tightly scoped to just the C stdlib for now (though the possibility of linking it with a dedicated [migration](https://github.com/conda-forge/conda-forge.github.io/issues/2015) for `error_overlinking: true` remains an option, if we want).
45
+
- (IF) Think we should have a mini-migrator (piggyback) so we don't have to rebuild all C/C++ packages; only rebuild when we really have to.
46
+
- (MvN) Agrees that we should take a lazy approach to migration; maybe have a list of some packages to eagerly migrate
47
+
- (MvN/IF) We can safely assume that the "world is flat" -> dependent packages won't need to be migrated as well.
- Recently we've migrated every 4th release of boost; previously it happened more often (every second release). There was a suggestion by a core member (Uwe) to migrate for 1.84; I think it'd be worth doing.
50
+
- After the big refactor with 1.82 (splitting off header-only lib, adding run-export), it should be easier to migrate nowadays.
51
+
- (IF) We should collect/review data on how long it takes to perform a boost migration, and use that to judge how often we should update. e.g. if it takes 3 months, then maybe once a year is sensible; if it takes 1 month, then every six months?
52
+
-[x] (KE) Can we create an sccache store to reduce build redundancy?
53
+
- (MvN) Big question is, "where do we put the cache?"
54
+
- (MCS) Do we have contacts at MSFT or other cloud providers we can talk with?
55
+
- (MvN) `conda-build` behavior complicates caching; e.g., use of timestamps in build env names can leak into cache if not careful
56
+
- (IF) When do we need sccache? E.g., does building different build numbers vs [upstream] versions benefit from cache?
57
+
- (MCS/KE) Opinion at Nvidia is Rapids can't get on conda-forge because it would take too long to build. Exploring if sccache would make conda-forge distribution feasible.
58
+
- (IF) using Quansight-hosted builder may be an option
59
+
- (KK) Building all of Rapids likely a bigger job than PyTorch or TensorFlow. May also want to consider splitting into per-device architecture builds
60
+
- (MCS) Need clear motivation to distribute Rapids via c-f; don't want to overload c-f infrastructure otherwise
61
+
- (KK) Currently not possible to [easily] depend on cuDF, cuML
0 commit comments