2026/03/07

Newest at the top

2026-03-07 12:31:40 +0100 <Guest89> actually, one thing I have been doing in general is guard patterns for unpacking. are they lazy as well or strict the same way pattern matching is?
2026-03-07 12:30:20 +0100 <Guest89> but it kind of feels like I am at the kitchen sink stage in general if I'm being honest
2026-03-07 12:29:55 +0100 <Guest89> I should probably experiment with it more but it seemed like allocations went down only a little (or at least less than I expected) while somehow runtimes increased slightly
2026-03-07 12:29:47 +0100Sgeo(~Sgeo@user/sgeo) (Read error: Connection reset by peer)
2026-03-07 12:29:16 +0100 <Guest89> I tried playing around with unboxed tuples and different pragmas like unpacking but they seemed to have varying/counterintuitive results
2026-03-07 12:27:58 +0100merijn(~merijn@host-cl.cgnat-g.v4.dfn.nl) (Ping timeout: 276 seconds)
2026-03-07 12:26:10 +0100wootehfoot(~wootehfoo@user/wootehfoot) (Read error: Connection reset by peer)
2026-03-07 12:23:55 +0100 <probie> If you need multiple return values and can't afford an allocation, look at unboxed tuples (although this is likely overkill)
2026-03-07 12:22:46 +0100 <probie> GHC is capable of doing something like `f x = (x, x+1)`, `g x = let (a, b) = f x in a + b` without actually allocating a tuple (assuming it can inline `f`), but that's the same for any user defined type as well
2026-03-07 12:22:18 +0100merijn(~merijn@host-cl.cgnat-g.v4.dfn.nl) merijn
2026-03-07 12:21:47 +0100 <Leary> Guest89: It might here and there, but that doesn't mean you're better off using them. In particular, they're lazy, so they can easily accumulate big thunks. I suggest replacing such parts of your representation with suitably strict bespoke data declarations.
2026-03-07 12:17:43 +0100 <Guest89> I just understood ghc optimized for tuples in particular over ordinary data constructors
2026-03-07 12:17:16 +0100 <probie> Why?
2026-03-07 12:17:07 +0100 <Guest89> I thought ghc treated them differently?
2026-03-07 12:16:09 +0100 <probie> Beyond the special syntax, tuples aren't really much different from `data T2 a b = T2 a b`, `data T3 a b c = T3 a b c`, `data T4 a b c d = T4 a b c d` etc.
2026-03-07 12:15:03 +0100merijn(~merijn@host-cl.cgnat-g.v4.dfn.nl) (Ping timeout: 265 seconds)
2026-03-07 12:11:08 +0100 <Guest89> seems the biggest allocations come from primitive types, tuples etc
2026-03-07 12:08:11 +0100merijn(~merijn@host-cl.cgnat-g.v4.dfn.nl) merijn
2026-03-07 12:02:52 +0100 <Guest89> also I've been relying on using eventlog2html but it seems to break fairly easily. are there any other options for visualizing the profiles?
2026-03-07 12:02:32 +0100ChaiTRex(~ChaiTRex@user/chaitrex) ChaiTRex
2026-03-07 12:01:49 +0100 <Guest89> I have some plots from using -hc that tells me which functions allocate the most but to be honest they're not particularly surprising in that department
2026-03-07 12:01:03 +0100 <Guest89> i'll give it a whirl
2026-03-07 12:00:51 +0100 <Guest89> sorry, -h
2026-03-07 12:00:49 +0100 <Leary> -hT: https://downloads.haskell.org/ghc/latest/docs/users_guide/profiling.html#rts-options-heap-prof
2026-03-07 11:59:50 +0100 <Guest89> it's one of the -l(x) RTS settings
2026-03-07 11:59:28 +0100 <haskellbridge> <sm> how do you do that Leary ?
2026-03-07 11:59:22 +0100hiecaq(~hiecaq@user/hiecaq) (Quit: ERC 5.6.0.30.1 (IRC client for GNU Emacs 30.2))
2026-03-07 11:59:04 +0100 <Guest89> my reference implementation generates only a few megabytes of data by comparison but again it's not comparable 1:1
2026-03-07 11:58:57 +0100CiaoSen(~Jura@2a02:8071:64e1:da0:5a47:caff:fe78:33db) CiaoSen
2026-03-07 11:58:49 +0100ChaiTRex(~ChaiTRex@user/chaitrex) (Ping timeout: 258 seconds)
2026-03-07 11:58:11 +0100 <Guest89> the only thing I haven't tried has to force computations in different places
2026-03-07 11:57:34 +0100 <Guest89> https://paste.tomsmeding.com/xZZPhSCR
2026-03-07 11:57:29 +0100Beowulf(florian@sleipnir.bandrate.org)
2026-03-07 11:57:14 +0100 <lambdabot> Help us help you: please paste full code, input and/or output at e.g. https://paste.tomsmeding.com
2026-03-07 11:57:14 +0100 <Leary> @where paste
2026-03-07 11:56:45 +0100 <Leary> Guest89: The problem is less likely to be allocations than unnecessary retention or unwanted thunks bloating your representation. Allocating is almost free, holding onto it is what costs you. In any case, I would start by heap profiling by type, which doesn't actually require a profiling build.
2026-03-07 11:56:37 +0100 <Guest89> I guess I'll just write it down:
2026-03-07 11:55:58 +0100 <Guest89> not familiar
2026-03-07 11:55:46 +0100arandombit(~arandombi@user/arandombit) (Ping timeout: 248 seconds)
2026-03-07 11:55:46 +0100 <haskellbridge> <sm> it's a lot easier if you use the matrix room
2026-03-07 11:55:31 +0100 <Guest89> or should I just put it in writing
2026-03-07 11:55:26 +0100 <Guest89> is it possible to paste pictures over IRC?
2026-03-07 11:55:14 +0100 <haskellbridge> <sm> well, how much memory does +RTS -s say is being allocated ?
2026-03-07 11:53:30 +0100 <Guest89> I just don't see how garbage collection can dominate the runtime so much
2026-03-07 11:52:49 +0100 <haskellbridge> <loonycyborg> ye it's just an example
2026-03-07 11:52:40 +0100 <Guest89> most of the actual allocation happens within the core algorithms, but I've also tried varying levels of strictness there
2026-03-07 11:52:17 +0100 <Guest89> I've tried both strict and non-strict versions of foldl, and I didn't seem to see any issues
2026-03-07 11:51:53 +0100 <haskellbridge> <sm> it requires investigation, there's no way round it
2026-03-07 11:51:37 +0100 <haskellbridge> <loonycyborg> Maybe it's thunk explosion somewhere, like from foldl
2026-03-07 11:51:26 +0100 <Guest89> obviously you can't get the performance *that* close, at least not in terms of memory, but as things are it's both an absurd difference and not feasible