B

80 posts

B banner
B

B

@BenCheesy

London, England Katılım Aralık 2012
37 Takip Edilen2 Takipçiler
B retweetledi
トレカパイク TORECA PIKE
🔥『ポケカ新弾 ロケット団の栄光』🔥 🎉🎉BOX抽選定価販売を致します🎉🎉 〜応募方法〜 1.トレカパイク(@Toreca_Pike )をフォロー 2.このポストをリポスト 〆切 4/17(水)23:59まで 当選者様にはこちらからdmにて ご連絡させて頂きます。 ✅支払い方法  ※郵送の方は銀行振り込みのみ ※店頭の方は現金のみ dmや店頭でのお問い合わせはお控え頂くようお願い致します🙇‍♂️ #トレカパイク #ポケモンカード #ポケカ #ロケット団の栄光 #BOX #Pokemon
トレカパイク TORECA PIKE tweet media
日本語
190
3.1K
1.3K
156.2K
B retweetledi
スニダン名古屋パルコ店(トレカ)【公式】
/ スニダン名古屋パルコ店トレカ取り扱いスタートOPEN記念🎉 「#テラスタルフェスex」ボックスを抽選定価販売 \ 📱応募方法 ①@snkrdunk_nagoyaをフォロー ②この投稿をいいね&リポスト 🗓応募締切 4/9(水)23:59 🎉当選発表 4/11(金)以降に@snkrdunk_nagoyaよりDMにてご連絡いたします ⚠️当選時にスニダン名古屋店にご来店いただける方のみご応募ください ⚠️受け取りはスニダン名古屋パルコ店のみになります
スニダン名古屋パルコ店(トレカ)【公式】 tweet media
日本語
102
3.4K
3.2K
170.1K
B retweetledi
カード王 大阪心斎橋🦇
【#ポケモンカードゲーム】#ポケカ 4月18日(金)発売‼️✨ 拡張パック #ロケット団の栄光 抽選販売のご案内です🔥 ①当店をフォロー ②このポストをリポスト 上記2点で応募完了✨ 詳細は画像をご確認下さい🙇🏻‍♂️✨
カード王 大阪心斎橋🦇 tweet media
日本語
185
3.3K
1.2K
247.1K
B retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
I was given early access to Grok 3 earlier today, making me I think one of the first few who could run a quick vibe check. Thinking ✅ First, Grok 3 clearly has an around state of the art thinking model ("Think" button) and did great out of the box on my Settler's of Catan question: "Create a board game webpage showing a hex grid, just like in the game Settlers of Catan. Each hex grid is numbered from 1..N, where N is the total number of hex tiles. Make it generic, so one can change the number of "rings" using a slider. For example in Catan the radius is 3 hexes. Single html page please." Few models get this right reliably. The top OpenAI thinking models (e.g. o1-pro, at $200/month) get it too, but all of DeepSeek-R1, Gemini 2.0 Flash Thinking, and Claude do not. ❌ It did not solve my "Emoji mystery" question where I give a smiling face with an attached message hidden inside Unicode variation selectors, even when I give a strong hint on how to decode it in the form of Rust code. The most progress I've seen is from DeepSeek-R1 which once partially decoded the message. ❓ It solved a few tic tac toe boards I gave it with a pretty nice/clean chain of thought (many SOTA models often fail these!). So I upped the difficulty and asked it to generate 3 "tricky" tic tac toe boards, which it failed on (generating nonsense boards / text), but then so did o1 pro. ✅ I uploaded GPT-2 paper. I asked a bunch of simple lookup questions, all worked great. Then asked to estimate the number of training flops it took to train GPT-2, with no searching. This is tricky because the number of tokens is not spelled out so it has to be partially estimated and partially calculated, stressing all of lookup, knowledge, and math. One example is 40GB of text ~= 40B characters ~= 40B bytes (assume ASCII) ~= 10B tokens (assume ~4 bytes/tok), at ~10 epochs ~= 100B token training run, at 1.5B params and with 2+4=6 flops/param/token, this is 100e9 X 1.5e9 X 6 ~= 1e21 FLOPs. Both Grok 3 and 4o fail this task, but Grok 3 with Thinking solves it great, while o1 pro (GPT thinking model) fails. I like that the model *will* attempt to solve the Riemann hypothesis when asked to, similar to DeepSeek-R1 but unlike many other models that give up instantly (o1-pro, Claude, Gemini 2.0 Flash Thinking) and simply say that it is a great unsolved problem. I had to stop it eventually because I felt a bit bad for it, but it showed courage and who knows, maybe one day... The impression overall I got here is that this is somewhere around o1-pro capability, and ahead of DeepSeek-R1, though of course we need actual, real evaluations to look at. DeepSearch Very neat offering that seems to combine something along the lines of what OpenAI / Perplexity call "Deep Research", together with thinking. Except instead of "Deep Research" it is "Deep Search" (sigh). Can produce high quality responses to various researchy / lookupy questions you could imagine have answers in article on the internet, e.g. a few I tried, which I stole from my recent search history on Perplexity, along with how it went: - ✅ "What's up with the upcoming Apple Launch? Any rumors?" - ✅ "Why is Palantir stock surging recently?" - ✅ "White Lotus 3 where was it filmed and is it the same team as Seasons 1 and 2?" - ✅ "What toothpaste does Bryan Johnson use?" - ❌ "Singles Inferno Season 4 cast where are they now?" - ❌ "What speech to text program has Simon Willison mentioned he's using?" ❌ I did find some sharp edges here. E.g. the model doesn't seem to like to reference X as a source by default, though you can explicitly ask it to. A few times I caught it hallucinating URLs that don't exist. A few times it said factual things that I think are incorrect and it didn't provide a citation for it (it probably doesn't exist). E.g. it told me that "Kim Jeong-su is still dating Kim Min-seol" of Singles Inferno Season 4, which surely is totally off, right? And when I asked it to create a report on the major LLM labs and their amount of total funding and estimate of employee count, it listed 12 major labs but not itself (xAI). The impression I get of DeepSearch is that it's approximately around Perplexity DeepResearch offering (which is great!), but not yet at the level of OpenAI's recently released "Deep Research", which still feels more thorough and reliable (though still nowhere perfect, e.g. it, too, quite incorrectly excludes xAI as a "major LLM labs" when I tried with it...). Random LLM "gotcha"s I tried a few more fun / random LLM gotcha queries I like to try now and then. Gotchas are queries that specifically on the easy side for humans but on the hard side for LLMs, so I was curious which of them Grok 3 makes progress on. ✅ Grok 3 knows there are 3 "r" in "strawberry", but then it also told me there are only 3 "L" in LOLLAPALOOZA. Turning on Thinking solves this. ✅ Grok 3 told me 9.11 > 9.9. (common with other LLMs too), but again, turning on Thinking solves it. ✅ Few simple puzzles worked ok even without thinking, e.g. *"Sally (a girl) has 3 brothers. Each brother has 2 sisters. How many sisters does Sally have?"*. E.g. GPT4o says 2 (incorrectly). ❌ Sadly the model's sense of humor does not appear to be obviously improved. This is a common LLM issue with humor capability and general mode collapse, famously, e.g. 90% of 1,008 outputs asking ChatGPT for joke were repetitions of the same 25 jokes​. Even when prompted in more detail away from simple pun territory (e.g. give me a standup), I'm not sure that it is state of the art humor. Example generated joke: "*Why did the chicken join a band? Because it had the drumsticks and wanted to be a cluck-star!*". In quick testing, thinking did not help, possibly it made it a bit worse. ❌ Model still appears to be just a bit too overly sensitive to "complex ethical issues", e.g. generated a 1 page essay basically refusing to answer whether it might be ethically justifiable to misgender someone if it meant saving 1 million people from dying. ❌ Simon Willison's "*Generate an SVG of a pelican riding a bicycle*". It stresses the LLMs ability to lay out many elements on a 2D grid, which is very difficult because the LLMs can't "see" like people do, so it's arranging things in the dark, in text. Marking as fail because these pelicans are qutie good but, but still a bit broken (see image and comparisons). Claude's are best, but imo I suspect they specifically targeted SVG capability during training. Summary. As far as a quick vibe check over ~2 hours this morning, Grok 3 + Thinking feels somewhere around the state of the art territory of OpenAI's strongest models (o1-pro, $200/month), and slightly better than DeepSeek-R1 and Gemini 2.0 Flash Thinking. Which is quite incredible considering that the team started from scratch ~1 year ago, this timescale to state of the art territory is unprecedented. Do also keep in mind the caveats - the models are stochastic and may give slightly different answers each time, and it is very early, so we'll have to wait for a lot more evaluations over a period of the next few days/weeks. The early LM arena results look quite encouraging indeed. For now, big congrats to the xAI team, they clearly have huge velocity and momentum and I am excited to add Grok 3 to my "LLM council" and hear what it thinks going forward.
Andrej Karpathy tweet media
English
668
2.2K
16.8K
3.7M
B
B@BenCheesy·
@alexkoh you’re the goat!
English
0
0
0
20
Alex Koh | BuyTrigger
Alex Koh | BuyTrigger@alexkoh·
Success in investing isn’t about buying every dip. It’s about holding onto strong, fundamentally sound stocks during tough times and riding the wave back up. But remember, be a two-way investor—protect your gains. Don’t be afraid to take profits and buy puts when everyone else is feeling confident.
English
3
0
18
2.9K
B
B@BenCheesy·
@alexkoh Thoughts on $UBER?
English
1
0
0
130
B
B@BenCheesy·
@alexkoh Still bullish on $HOOD?
English
0
0
0
110
Alex Koh | BuyTrigger
Alex Koh | BuyTrigger@alexkoh·
To all my 🇺🇸 friends, make sure to go out and vote! May the best candidate win. For my fellow family investors: be prepared for volatility next week. ✅ Last chance to BTFD ✅ Final portfolio adjustments for 2025 ✅ Lock in that growth portfolio As Sun Tzu says, “In times of chaos, there is always opportunity.”
English
1
0
7
813
B
B@BenCheesy·
@alexkoh current buy triggers for $HOOD?
English
0
0
0
45
Alex Koh | BuyTrigger
Alex Koh | BuyTrigger@alexkoh·
Back in 7th May I called All Time High. Six weeks later it’s still going higher and close to doubling last ATH year to date. I tried to swing for profits but it looks like the holding (5️⃣🐎) and DCA on selective stocks have helped. I am sure there are others with 2x going to 4x for YTD by now. Well done all! My 5 🐎 $NVDA $AVGO $TSM $UBER $HOOD 🚀 I missed $QCOM $SMCI $ARM and I am not perfect. If I said I am not worried, that is an absolute lie 😂! Brainstorming my next action. Will share video when ready.
Alex Koh | BuyTrigger tweet media
Alex Koh | BuyTrigger@alexkoh

I am up close to 35% and still have plenty of cash. I did by the dip on that two little blip and happy. This is a growth diversification. I would like focused AI portfolio but my ❤️ cannot handle the volatility. With volatility focused portfolio you must trade to hedge. I don’t have the time. $NVDA and $UBER will pave ways to all time high if they succeed. Remember we are on a Sideway market and reason I am bouncing is BTFD. Don’t be afraid on volatility cause your homework will pave way. Have a good day. Just one post from me today! Your family friend Alex!

English
1
0
13
1.7K
B
B@BenCheesy·
@alexkoh $150 short term bounce for an options play. if it doesn’t break $180 (which it wont), we will see $130-140 next
English
1
0
1
110
Chicken Genius
Chicken Genius@pakpakchicken·
good dip. . you scared? 😂 added some positions I missed. bitches we are in a bull market. Market dip due to fear of war. After war starts, market doesn't give a damm. Haven't you learn? Of course, not asking you to 10X leverage full port.
English
120
78
1.5K
173.5K
B
B@BenCheesy·
@alexkoh Dr Alex, care to share your thoughts on $MOMO?
English
1
0
1
82
Alex Koh | BuyTrigger
Alex Koh | BuyTrigger@alexkoh·
New YouTube Video [Part 1] is out. In 2024, systematic investing is the way to go. Engineer and investor, Alex Koh, shares his top 22 stock picks for the year. From staple stocks like Apple and Microsoft to high-growth investments like AI companies and Uber, Koh breaks down his selections based on revenue growth, long-term investment potential, and macroeconomic impact. He emphasizes the importance of data analysis and a long-term view for building wealth and securing future financial stability for family investors. Watch the full video for in-depth insights on these stock picks. youtu.be/N7G-KbiAOo8
YouTube video
YouTube
Alex Koh | BuyTrigger@alexkoh

Sharing my video soon. I have 22 stocks on my shortlist or already invested in my family investment portfolio. My super editor Mosh is putting in the final edits to tweak it up. Warning it’s longer than usual video but it allows me to explain my views. I am late but at least I have latest earnings data update on it. Next video would be my “How to Buy Trigger”. A lot of work to be done and update on my free website. Hope to serve and share values. PS: is the thumbnail better than my old boring 💩 😂 #familyinvestments #father3jobs

English
1
0
4
1.7K
B
B@BenCheesy·
@legit_demon go away I'm studying for exams
English
0
0
0
0