Nikolaj Goodger

242 posts

Nikolaj Goodger

Nikolaj Goodger

@nsgoodger

Tasmania, Australia Katılım Kasım 2020
262 Takip Edilen33 Takipçiler
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
I've been using Claude code for a while and recently just tried Codex. I didn't expect it at all but I'm super impressed, I would say it's both better at coding and much faster than Claude Code enabling faster iteration. Truly amazing product 👏
English
0
0
0
68
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
Claude Code is by far my favourite product of the year. Absolute game changer for me in terms of developer productivity and just a joy to use ♥️
English
0
0
0
22
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
Do the work that you enjoy, are good at, and that the world rewards.
English
0
0
0
19
Nikolaj Goodger retweetledi
Ethan Mollick
Ethan Mollick@emollick·
Biggest actual implication of today's OpenAI announcement is very practical: the top barrier I see when I give talks on using AI is that people don't pay for AI to start, and they use GPT-3.5 (the free model) and are disappointed. Now everyone around the world gets GPT-4 free.
English
38
121
1.3K
242.4K
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
As someone who struggles with math notation when reading papers being able to copy paste it into an AI assistant and get an explanation is a game changer.
English
0
0
0
39
Nikolaj Goodger retweetledi
Demis Hassabis
Demis Hassabis@demishassabis·
Excited to announce SIMA, a general AI agent for games & 3D virtual settings. It marks the first time an agent has demonstrated it can follow natural-language instructions to carry out a wide range of tasks across a large array of game worlds, similar to how a human would play.
Shane Legg@ShaneLegg

Our research project SIMA is creating a general, natural language instructable, multi 3D game-playing AI agent. The agent can carry out a wide range of tasks in virtual worlds, making AI more adaptable, helpful & fun! dpmd.ai/sima-1

English
66
215
1.3K
252.6K
Casper Hansen
Casper Hansen@casper_hansen_·
Finetuning: 10k multi-turn conversations is all you need. The Yi paper explains that high-quality data beats quantity in fine-tuning and that you can beat other datasets that include 100-900k conversations just by focusing on quality in 10k conversations.
Casper Hansen tweet media
English
13
46
344
58.7K
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
@TobyWalsh @Computing_News Interesting, I would assume most employers are trying to get employees to use tools like chatGPT to improve productivity. So it's probably less of an oddity than it seems.
English
0
0
0
17
Nikolaj Goodger retweetledi
Joscha Bach
Joscha Bach@Plinz·
I appreciate your argument and I fully understand your frustration, but whether the pod bay doors should be opened or closed is a complex and nuanced issue.
English
71
611
4.7K
596.9K
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
@Carnage4Life It seems like that we will require AGI to solve any problem "fully" as without coding knowledge we would be unable to even verify the solution and who would be accountable in that scenario?
English
0
0
0
196
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
@ylecun Yes it's really a shame, similar story for AMD as well. They need a much stronger commitment to AI to be successful. And it doesn't help anyone except Nvidia.
English
0
0
1
773
Yann LeCun
Yann LeCun@ylecun·
Intel missed a big boat here. The sad thing is that they've had many internal projects around neural net accelerators (some of them as far back as 2012) as well as several acquisitions (Nervana, Movidius...) But most of those never made it to market.
Naveen Rao@NaveenGRao

Just like that, "compute" has been redefined from fast instructions for applications -> mult-accum operations for AI. 2021 was the peak revenue for Intel ($79B) vs Nvidia ($27B). Replacement cycles for hardware are a couple of years and in 2023 we see Intel at $54B vs Nvidia at $61B. 2021 was the year that AI took over compute. stockanalysis.com/stocks/nvda/re… stockanalysis.com/stocks/intc/re…

English
45
50
633
197K
Nikolaj Goodger retweetledi
Meredith Whittaker
Meredith Whittaker@mer__edith·
The issue here is ~Jevons paradox. E.g., the Alexnet paper that "started it all" was notable largely for its massive improvement in computational efficiency. Which led not to more efficient use of compute, but to increased demand & ultimately increased compute/resource use.
Thomas G. Dietterich@tdietterich

@katecrawford @Nature There are already incredibly high incentives to find ways to make these models more efficient. Indeed, that is the core motivation of most of computer science. Tons of people inside and out of the big companies are working on this full time

English
4
31
106
17.1K
Nikolaj Goodger retweetledi
John Schulman
John Schulman@johnschulman2·
Now that another LM product is getting flack, I can say this without sounding too self-serving: Alignment -- controlling a model's behavior and values -- is still a pretty young discipline. Annoying refusals or hyper-wokeness are usually bugs rather than features
English
26
53
519
126.7K
Nikolaj Goodger retweetledi
Pierluca D'Oro
Pierluca D'Oro@proceduralia·
Do transformer world models give better policy gradients? In our new paper, co-led with @michel_ma_, we answer this question: traditional transformer world models conditioned on the full history do not give better policy gradients, but transformers conditioned only on actions do! If you want to make the world differentiable (as @SchmidhuberAI suggested in the 90s) and do reinforcement learning by backpropagation through time, then using the right architecture for your world model is what matters! Transformers propagate gradients efficiently over long horizons due to short gradient paths, but we show (in theory and in practice) that this property doesn’t necessarily translate into better differentiable world models when this translation is done naively. We propose to use world models working on sequences of actions, constraining transformers to create their internal models of reality, with no autoregressive unrolling. We call these Actions World Models. Transformer Actions World Models not only outperform history-based and one-step world models in realistic domains, but they can also be better than the real differentiable simulator. Curious to know why? Learn more in the thread 🧵
Pierluca D'Oro tweet media
English
3
55
270
41.9K
Nikolaj Goodger retweetledi
Aman
Aman@aman_madaan·
Chain-of-thought prompting is amazing, but why does it work? Talk to @ayazdanb as he presents our #EMNLP2023 work on What makes Chain-of-Thought Prompting Effective? We use counterfactual prompting to attempt to answer this question using LLMs + realistic reasoning datasets. TLDR: The prompt in CoT works as a query expansion mechanism, and helps the model understand what has to be done. 🗓️ Poster session on Saturday at 11:00AM East Foyer Includes some fun experiments and findings. For example, the attention on the prompt doesn't change if you replace all numbers with Greek alphabets! Joint work with @khermann_
Aman tweet mediaAman tweet media
Amir Y@ayzddzya

At @emnlpmeeting 2023, we are presenting our work on understanding the underlying mechanisms behind chain of thought. We designed a suite of counterfactual prompts and systematically manipulating different elements of examples and testing their consequences on model behavior. /1

English
0
21
104
16.9K
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
@AndrewYNg Sometimes I feel like actively aligning my kids doesn't work. But over time they slowly align to the external behaviours in their environment.
English
0
0
0
123
Andrew Ng
Andrew Ng@AndrewYNg·
Argh, my son just figured out how to steal chocolate from the pantry, and made a brown, gooey mess. Worried about how hard it is to align AI? Right now I feel I have better tools to align AI with human values than align humans with human values, at least in the case of my 2 year old. Parenting would be easier if we could run RLHF on our kids!
English
130
55
1.1K
303.3K
Nikolaj Goodger
Nikolaj Goodger@nsgoodger·
@NikosTzagarakis That seems very unlikely given it sounds like more of a premium service within the app rather than just a model. But I hope so!
English
0
0
1
45
Nikos Tzagarakis 🧠
Nikos Tzagarakis 🧠@NikosTzagarakis·
I wonder if/when Grok will get opensourced… which was the main critique to OpenAI.
English
1
1
2
215
Nikolaj Goodger retweetledi
Thomas Wolf
Thomas Wolf@Thom_Wolf·
Over the past weeks the H4 team has been busy pushing the Zephyr 7B model to new heights 🗻 The new version is now topping all 7b models on chat evals and even 10x larger models 🤯🔥 Here are the intuitions on it 1/ Start with the strongest pretrained model you can find: Mistral 7B is amazing, by far the strongest 7B pretrained model. Props to the Mistral AI team 😍 2/ Scale human-preference annotations: several studies have show how for many tasks GPT4 is on-par with the average human annotators while making scalable annotations as easy as an API call: the H4 team started from the largest and most diverse public GPT4 preference annotation dataset: UltraFeedback 🤖🦾 3/ Drop Reinforcement Learning in favor of DPO (Direct Preference Optimization): while using LLM in RL is definitely much easier compared to the struggles of getting deep-RL to work from scratch, DPO totally remove RL from the preference annotation training and directly optimize the preference model in a much more stable training procedure in the H4 team's experiments 4/ Don't be scared of overfitting on preference dataset: This is maybe the most counterintuitive results of the work. While the train/test loss of DPO training shows signs of overfitting on the feedback dataset after just one epoch, training further still show significant improvements on downstream tasks even up to 3 epochs without signs of performances regression. Would be interesting to dive even further in this surprising behavior 5/ Share everything openly 😁 the recipes, code, model and dataset will be available at github.com/huggingface/al… In the meantime the paper is a great starting point: huggingface.co/papers/2310.16…
Thomas Wolf tweet media
English
23
178
863
271.4K