lots of people offering jobs over there
but conversation is much better over here
lots of people offering jobs over there
but conversation is much better over here
get grinding with claude asap, build that muscle
We need to raise the bar on research code right now.
1) documentation and tests are dead simple now.
2) creating benchmarks integrating across multiple implementations
3) have agents double check your work / fix broken tests
4) fix outstanding bugs in major scientific packages
@hempuli.bsky.social !
We want to expand on utilForever's Baba is You RL simulator!
github.com/utilForever/...
Is that ok with you? We want to expand the set of levels included to involve more complex problems and reasoning over the various abstractions in your (awesome and very hard) game!
Massive shoutout to the efforts of Sanghyeok Choi (he's on the dark MAGA app), Salem Lahlou (mbzuai.ac.ae/study/facult...), and βͺβͺ@oyounis.bsky.social - this was very much a team effort and weβre really excited to help popularize gflownet use through these tools. We really value your feedback!
If youβre interested in using torchgfn, helping us improve the library, want help incorporating torchgfn into your workflow, or have any feedback, please feel free to familiarize yourself with our documentation and reach out β thereβs still lots to do!
torchgfn.readthedocs.io/en/latest/
Moving forward, we plan to focus on optimizing the library for large-scale distributed training setups, and supporting more specialized and demanding environments, particularly in the AI for Science domains.
The structure of the GFlowNet itself is highly modular, permitting the use of modified losses, custom samplers, novel off-policy sampling methods, and new policy architectures with minimal changes to the underlying library elements.
Basic torchgfn usage follows standard pytorch workflows, allowing the user to swap in any modified components to support the development of new methods:
This is a major update:
+ Much easier environment definition.
+ Cleaner abstractions β easier extensibility!
+ Support for graph-based states under torch_geometric.
+ Improvements to every core element of the library.
+ Lots of new environments, tutorials, and examples!
Weβve released torchgfn v2!
github.com/GFNOrg/torch...
We believe weβve built the go-to library for fundamental GFlowNet methods development and prototyping, and weβre really excited to help you start using it.
Today marks a big milestone for me. I'm launching @law-zero.bsky.social, a nonprofit focusing on a new safe-by-design approach to AI that could both accelerate scientific discovery and provide a safeguard against the dangers of agentic AI.
Yeah the lack of memepoasters and tpot adjacent attention bait makes the platform great for us but bad for it taking over the disinformation psyop town square.
America cannot long remain free, nor first among nations, if it becomes the kind of place where universities are dismantled because they don't align politically with the current head of the government.
Preprint Alert π
Can we simultaneously learn transformation-invariant and transformation-equivariant representations with self-supervised learning?
TL;DR Yes! This is possible via simple predictive learning & architectural inductive biases β without extra loss terms and predictors!
π§΅ (1/10)
Great job gang! And thanks to @boussifo.bsky.social for being such a stellar lead. (Also depicted: @jainmoksh.bsky.social)
@boussifo.bsky.social , LΓ©na NΓ©hale Ezzine, MichaΕ Koziarski, @jainmoksh.bsky.social , Nikolay Malkin, @bengioe.bsky.social , Rim Assouel, @yoshuabengio.bsky.social - @mila-quebec.bsky.social
Action Abstractions for Amortized Sampling
π arxiv.org/abs/2410.15184
π» github.com/GFNOrg/Chunk...
The additive benefit of combining chunking with diversity-seeking samplers, like GFlowNets, also points towards an intriguing explanation as to why macro action discovery has not been found generally useful in the RL context.
π‘ Why is this exciting?
Hierarchical planning is a key component of intelligenceβboth biological & artificial. By dynamically learning & using abstractions, our method bridges the gap between RL, program induction, and cognitive science.
These chunks also generalize - theyβre transferable across samplers and tasks!
Chunks learned in one environment improve exploration and sampling efficiency in unseen settings, suggesting the method abstracts high order general principles that are robust & adaptable to new envs!
For mode discovery, our approach also significantly speeds up discovering diverse high-reward states.
For example, in FractalGrid, vanilla GFlowNets get stuck in a single mode, but armed with ActionPiece, it unlocks new exploration paths!
Chunking helps!
Across synthetic and real-world tasks (e.g., RNA sequence generation, bit sequences, and FractalGrid), our approach improves especially for GFlowNets:
β
Mode discovery
β
Exploration
β
Density estimation
β
Interpretability
By applying BPE (which we're calling "ActionPiece" for learning chunked actions) to sampled trajectories, we extract meaningful high-level actions that naturally emerge during learning. For example, here are some learned chunks from sampler of RNA binders:
We chunk frequently occurring subsequences into high-order actions using Byte Pair Encoding (BPE)βa popular NLP tokenization technique. These chunks are added to the action space, which progressively reduces trajectory length and helps uncover latent structures in the task.
In RL & GFlowNets, with longer trajectories, assigning credit and discovering diverse high-reward states gets harder. Standard methods struggle to sample structured distributions efficiently & many previous attempts to discover high-order actions failed to show consistent benefit.
Ecstatic to show off some work my brilliant colleagues and I did at @iclr-conf.bsky.social this year! π
We address the credit assignment challenge under long trajectories in RL or GFlowNets by constructing high order actions, or βchunksβ, effectively compressing trajectory lengths!
From having to wrap a bunch of methods I can say I'm not a big fan of having configs in the form of argparsers.
Musk tweet: CFPB RIP
any economists here that can help me understand why the world's richest man might want to kill the consumer financial protection bureau
Neural surrogates can accelerate PDE solving but need expensive ground-truth training data. Can we reduce the training data size with active learning (AL)? In our NeurIPS D3S3 poster, we introduce AL4PDE, an extensible AL benchmark for autoregressive neural PDE solvers. π§΅
"The complaint alleges the algorithm, dubbed nH Predict, has a 90% error rate, basing that calculation on the percentage of payment denials reversed through internal appeals processes or administrative law judge rulings."
www.statnews.com/2023/11/14/u...