Jiri Simsa

145 posts

Jiri Simsa banner
Jiri Simsa

Jiri Simsa

@jsimsa

Working on data processing and analysis infrastructure for ML @ Google.

California, USA Katılım Eylül 2015
0 Takip Edilen165 Takipçiler
Jiri Simsa
Jiri Simsa@jsimsa·
If you are interested in advancing infrastructure that provides large scale data analysis and processing for ML workloads across Google, my team is hiring: linkedin.com/jobs/view/2905…
English
0
8
7
0
Jiri Simsa retweetledi
Google
Google@Google·
Five years ago, we open sourced @TensorFlow, our machine learning framework that's now the most popular machine learning library in the world. 🌎 To celebrate, we’re sharing few interactive demos and tutorials you can try, no experience required → goo.gle/3nz22Xh
GIF
English
22
455
2.3K
0
Jiri Simsa retweetledi
James Bradbury
James Bradbury@jekbradbury·
In 2016, when I was working on machine translation, it took me more than a week on a multi-GPU machine to train a competitive system on WMT English-German. Today, JAX on a TPU v3 supercomputer can train a better model on the same data in 16 seconds! cloud.google.com/blog/products/…
James Bradbury tweet media
English
10
139
870
0
Jiri Simsa retweetledi
👩‍💻 Paige Bailey
👩‍💻 Paige Bailey@DynamicWebPaige·
👉 tf.data supports *any* machine learning framework (JAX, @TensorFlow, PyTorch, more!), and is a great way to speed up your data input pipelines. Be sure to try out our new features for tf.data, available in TF 2.3: #diff-781a53e648f3df8d16a08ec083b04bf4" target="_blank" rel="nofollow noopener">github.com/tensorflow/ten…
Ong Chin Hwee 🐼@ongchinhwee

1. Start with TF Data 2. Enable non-deterministic ordering 3. Cache data 4. Turn on experimental optimizations 5. Autotune parameter values --> >10% performance improvement! 🤯 #EuroPython

English
6
45
156
0
Jiri Simsa retweetledi
TensorFlow
TensorFlow@TensorFlow·
🔍Inside TensorFlow: tf.data + tf.distribute In this presentation, Jiri Simsa showcases best practices. You’ll learn about the input pipeline, parallel extraction, distributed training, and more. Watch here → goo.gle/2wYGEG7
TensorFlow tweet media
English
1
33
123
0
Jiri Simsa retweetledi
Josh Gordon
Josh Gordon@random_forests·
If your dataset is small, use an in-memory cache: ds = ds.cache() If large, create an on-disk cache: ds = ds.cache("my_file") Afterwards, you can call ds.batch() and ds.shuffle() as always. Complete example: tensorflow.org/tutorials/load…
English
4
22
175
0
Jiri Simsa retweetledi
Jeff Dean
Jeff Dean@JeffDean·
Not only are TPUs fast for doing machine learning, but they are also more energy efficient than alternative platforms, so you can feel great as you train that language model on scientific articles about climate change. twitter.com/GCPcloud/statu…
Google Cloud Tech@GoogleCloudTech

Our Cloud TPUs are designed with energy efficiency in mind, specifically to accelerate deep learning workloads at higher teraflops per watt compared to general purpose processors → blog.google/topics/google-… #EarthDay

English
4
82
381
0
Jiri Simsa retweetledi
Brennan Saeta
Brennan Saeta@bsaeta·
Today in #CloudTPU announcements: (1) @TensorFlow 1.8 now available with a slew of perf improvements (2.7k to 3.2k images/sec on ResNet-50, aka 12.5 hours is now 9 hours to fully train), and (2) we have opened up a new zone (us-central1-b) for HA & load balancing.
English
0
20
37
0
Jiri Simsa retweetledi
Jeff Dean
Jeff Dean@JeffDean·
We just posted new DAWNBench results for ImageNet classification training time and cost using Google Cloud TPUs+AmoebaNet (architecture learned via evolutionary search). You can train a model to 93% top-5 accuracy in <7.5 hours for <$50. Results: dawn.cs.stanford.edu/benchmark/
Jeff Dean tweet media
English
6
203
551
0