Felix Wick

348 posts

Felix Wick

Felix Wick

@WickFelix

Katılım Mayıs 2014
6 Takip Edilen278 Takipçiler
Felix Wick
Felix Wick@WickFelix·
Try it out for your probabilistic, ML-based decision making!
English
0
0
0
130
Felix Wick
Felix Wick@WickFelix·
In our latest Cyclic Boosting release github.com/Blue-Yonder-OS…, we introduce quantile regression (via pinball loss) and subsequent (requiring 3 predicted quantiles) fit-less estimation of full individual probability distributions by means of quantile-parameterized distributions.
English
1
2
6
925
Felix Wick
Felix Wick@WickFelix·
brief summary of main ML concepts: @felixwick83/understanding-machine-learning-96fd6280e2eb" target="_blank" rel="nofollow noopener">medium.com/@felixwick83/u…
Felix Wick tweet media
English
0
1
2
274
Felix Wick
Felix Wick@WickFelix·
First open-source pre-release (still lots of polishing needed) of the Cyclic Boosting ML algorithms: github.com/Blue-Yonder-OS… Feel free to try it out, simply do: pip install cyclic-boosting Please come back with criticism and suggestions. Contributors highly welcome!
English
0
2
12
603
Felix Wick retweetledi
The Cultural Tutor
The Cultural Tutor@culturaltutor·
A short history of the world in 13 maps & infographics: 1. Everybody alive today compared to everybody who has ever lived
The Cultural Tutor tweet media
English
262
11.3K
84.8K
0
Felix Wick retweetledi
Don Winslow
Don Winslow@donwinslow·
This is legit ninja shit.
English
891
11K
110.9K
0
Felix Wick
Felix Wick@WickFelix·
@antgoldbloom For products and locations, you need to find a way to deal with categorical features of high cardinality though. Simple one-hot encoding is not great and especially tree-based methods suffer here.
English
0
0
1
0
Felix Wick
Felix Wick@WickFelix·
@antgoldbloom A single model used across all SKUs is not only better in terms of learning commonalities across SKUs, but also operationally way more convenient.
English
1
0
2
0
Anthony Goldbloom
Anthony Goldbloom@antgoldbloom·
Been spending the last few weeks speaking to data scientists working on demand forecasting. Some interesting things I learned. 🧵
English
10
44
276
0
Felix Wick
Felix Wick@WickFelix·
@antgoldbloom One of the most important, but usually overlooked, issues (especially for short-term forecasting) is temporal confounding. Autocorrelation is spurious and can mask true causal effects, e.g., from promotions or events, for the model. Take care how you include it.
English
0
0
4
0
Felix Wick
Felix Wick@WickFelix·
@antgoldbloom This is better in terms of variance, but you pay for it with bias. In general, best is to learn on the granularity you want to predict.
English
1
0
0
0
Anthony Goldbloom
Anthony Goldbloom@antgoldbloom·
More mature teams with many SKUs are typically only forecasting demand at the most zoomed out level (e.g. the national level). And then forecast "share of sales" at more granular geographic levels.
English
2
0
14
0
Felix Wick
Felix Wick@WickFelix·
@PengM83 @antgoldbloom For new products (and locations) use attributes (including product groups if hierarchy available) and go for embeddings.
English
0
0
0
0
Peng
Peng@PengM83·
@antgoldbloom what about totally new product with no previous history? I've seen people talking about creating product cluster and then using prediction from similar products. Any real world observations on that? Thanks.
English
2
0
0
0
Mark Tenenholtz
Mark Tenenholtz@marktenenholtz·
Feature engineering is the most important part of building great models for tabular data. However, it’s easy to run out of ideas. Much like Writer’s Block, I call this Feature Engineering Block. So here are a bunch of ideas to make sure you never run out:
English
29
144
889
0
Felix Wick
Felix Wick@WickFelix·
@marktenenholtz Two more options: Learn as a distinct category or ignore by setting to a neutral value (if your algorithm allows that).
English
0
0
0
0
Mark Tenenholtz
Mark Tenenholtz@marktenenholtz·
Always start by filling them with a value that makes sense. If there isn’t an obvious value, try: • Filling with the mean • Filling with the median • Filling with zero For categorical features: • Filling with the mode • Filling with a negative value • Frequency encoding
English
2
2
35
0
Peter Hoffmann
Peter Hoffmann@peterhoffmann·
Got **the** book. Looking forward to learn quite some new stuff out of my data engineering comfort zone. @rasbt
Peter Hoffmann tweet media
English
13
34
347
0
Felix Wick
Felix Wick@WickFelix·
Was it really just symmetry breaking via setting random initial weights what Rosenblatt missed to get backpropagation to work? (as described in Genius Makers)
English
0
0
0
0