SK

1.2K posts

SK banner
SK

SK

@hey2SK

My linktr: https://t.co/NW7QpsNOUr Our webapp: https://t.co/R40oAWtBEo, New product: One-Click Explainer Video Creation

Katılım Ağustos 2023
288 Takip Edilen142 Takipçiler
Sabitlenmiş Tweet
SK
SK@hey2SK·
book in the 1am. its worth to read.
English
2
0
2
703
李老师不是你老师
李老师不是你老师@whyyoutouzhele·
马斯克小儿子穿中国风马甲 5月14日上午,马斯克与苹果CEO库克、英伟达CEO黄仁勋等十余名美方商界代表一同进入中美元首会谈现场。 引人注目的是,54岁的马斯克此行带上了6岁的小儿子,照片显示他穿着一件带有中式元素的上衣。
李老师不是你老师 tweet media
中文
809
1.4K
24.4K
3.4M
SK retweetledi
白水 桃花|nemuiasa
大家好,我來自日本 👋🏻 做了一款 app,叫「hibi」 一天那麼長, 最後濃縮成兩個字。 也許是「疲憊」,也許是「微光」, 也許只是「還好」—— 兩個字,卻什麼都說到了。 不用寫完整的句子, 不用交代來龍去脈, 漢字自己會呼吸。 原是為日文使用者而做, 沒想到華語的朋友們也悄悄找來了。 繁體中文圈的你, 今天,是哪兩個字?
白水 桃花|nemuiasa tweet media白水 桃花|nemuiasa tweet media
中文
62
34
484
54.6K
SK
SK@hey2SK·
Video upscaler for Chinese AI short drama.
SK tweet media
English
0
0
0
22
SK retweetledi
Tongyi Lab
Tongyi Lab@Ali_TongyiLab·
Meet Z-Anime by @seesee. Built on the powerful Z-Image Base architecture, this model brings flagship-level diversity and precise prompt control to anime generation. It inherits full support for complex negative prompts and extreme customizability. Ready to push your anime workflows to the next level? Grab the weights on Hugging Face:👇 huggingface.co/SeeSee21/Z-Ani…
Tongyi Lab tweet mediaTongyi Lab tweet media
English
8
47
419
72.6K
SK
SK@hey2SK·
RIght now learning infra immediately.
English
0
0
0
2
SK
SK@hey2SK·
my school days, I preferred geometry over algebra; visual graphics were more effective than logic for me.
English
0
0
0
6
SK
SK@hey2SK·
To CV engineer and builder, I always agree that visualized content is higher than symbolic system.
English
0
0
0
8
Yuchen Jin
Yuchen Jin@Yuchenj_UW·
When I want to learn something new, or dig into a paper, I have Claude generate a HTML for me. This works surprisingly well (especially in Claude, since Codex generated HTML is still kinda ugly...) It's better than Google NotebookLM. Podcasts are nice, but reading is much higher-bandwidth than listening to a podcast. HTML has a key advantage: they can show things. Diagrams. Charts. Interactive bits. You can actually poke at the idea, not just passively consume it. Then I iterate. Ask questions. Refine sections. Add missing pieces. The HTML evolves with my understanding. Over time, this compounds into a personal knowledge base. "The input/output mind meld between humans and AIs is ongoing and there is a lot of work to do and significant progress to be made, way before jumping all the way into neuralink-esque BCIs and all that." 💯
Andrej Karpathy@karpathy

This works really well btw, at the end of your query ask your LLM to "structure your response as HTML", then view the generated file in your browser. I've also had some success asking the LLM to present its output as slideshows, etc. More generally, imo audio is the human-preferred input to AIs but vision (images/animations/video) is the preferred output from them. Around a ~third of our brains are a massively parallel processor dedicated to vision, it is the 10-lane superhighway of information into brain. As AI improves, I think we'll see a progression that takes advantage: 1) raw text (hard/effortful to read) 2) markdown (bold, italic, headings, tables, a bit easier on the eyes) <-- current default 3) HTML (still procedural with underlying code, but a lot more flexibility on the graphics, layout, even interactivity) <-- early but forming new good default ...4,5,6,... n) interactive neural videos/simulations Imo the extrapolation (though the technology doesn't exist just yet) ends in some kind of interactive videos generated directly by a diffusion neural net. Many open questions as to how exact/procedural "Software 1.0" artifacts (e.g. interactive simulations) may be woven together with neural artifacts (diffusion grids), but generally something in the direction of the recently viral x.com/zan2434/status… There are also improvements necessary and pending at the input. Audio nor text nor video alone are not enough, e.g. I feel a need to point/gesture to things on the screen, similar to all the things you would do with a person physically next to you and your computer screen. TLDR The input/output mind meld between humans and AIs is ongoing and there is a lot of work to do and significant progress to be made, way before jumping all the way into neuralink-esque BCIs and all that. For what's worth exploring at the current stage, hot tip try ask for HTML.

English
32
24
421
43.5K
SK
SK@hey2SK·
I agree with it, and now my agent has accepted this purpose, and its medium text assets all used HTML. Why not, because markdown only is compatibe with some people who is willing to make friends with computer rather than human.
Andrej Karpathy@karpathy

This works really well btw, at the end of your query ask your LLM to "structure your response as HTML", then view the generated file in your browser. I've also had some success asking the LLM to present its output as slideshows, etc. More generally, imo audio is the human-preferred input to AIs but vision (images/animations/video) is the preferred output from them. Around a ~third of our brains are a massively parallel processor dedicated to vision, it is the 10-lane superhighway of information into brain. As AI improves, I think we'll see a progression that takes advantage: 1) raw text (hard/effortful to read) 2) markdown (bold, italic, headings, tables, a bit easier on the eyes) <-- current default 3) HTML (still procedural with underlying code, but a lot more flexibility on the graphics, layout, even interactivity) <-- early but forming new good default ...4,5,6,... n) interactive neural videos/simulations Imo the extrapolation (though the technology doesn't exist just yet) ends in some kind of interactive videos generated directly by a diffusion neural net. Many open questions as to how exact/procedural "Software 1.0" artifacts (e.g. interactive simulations) may be woven together with neural artifacts (diffusion grids), but generally something in the direction of the recently viral x.com/zan2434/status… There are also improvements necessary and pending at the input. Audio nor text nor video alone are not enough, e.g. I feel a need to point/gesture to things on the screen, similar to all the things you would do with a person physically next to you and your computer screen. TLDR The input/output mind meld between humans and AIs is ongoing and there is a lot of work to do and significant progress to be made, way before jumping all the way into neuralink-esque BCIs and all that. For what's worth exploring at the current stage, hot tip try ask for HTML.

English
0
0
0
6
SK
SK@hey2SK·
GCP AI agent service is bullshit.
English
0
0
0
7
SK
SK@hey2SK·
To ai videos, 24/30FPS, VFI model is a huge advantages to them. Like BGM, people always like 120/180 BGM.
English
0
0
0
14
SK
SK@hey2SK·
Making a whole landing page with the logon state, only use codex computer use and a little of skills. All functions have been verified to be working properly. Fuckkkkkkkk~, A powerful knives in my hand. vibe building is so fast.🚀🚀🚀
SK tweet media
English
0
1
2
32
SK
SK@hey2SK·
@bdsqlsz powerful force
English
0
0
0
99
青龍聖者
青龍聖者@bdsqlsz·
Codex is the winner; now I can modify the code and use Twitter at the same time while waiting for GPT5.5 ​coding.
青龍聖者 tweet media
English
1
0
24
3.9K
SK
SK@hey2SK·
SK tweet media
ZXX
0
0
0
9
SK
SK@hey2SK·
Who wouldn't love it?❤️❤️❤️
SK tweet media
English
1
0
2
10
SK retweetledi
GENEL | AIを用いた動画制作
ChatGPT Images 2.0 プロンプト: まるで小さな子供がクレヨンで描き直したような、やさしくて少し不器用な絵本風のイラスト。 人物の顔や手、服の細かい部分は描き込みすぎず、丸く柔らかい形で単純化する。 表情はあたたかく、少しいたずらっぽく、友達との楽しい記憶がにじむように描く。 色は実際の色に縛られず、明るいパステルカラーを自由に使う。 ピンク、ミント、レモンイエロー、ラベンダー、水色などをランダムに混ぜ、子供らしい想像力で再解釈する。 線はかなり揺れていて、塗りはクレヨン特有のざらざらした質感、色むら、にじみ、紙にこすれたような跡を感じさせる。 子供が大切な思い出を楽しく飾ったような自然な配置にする。 全体は子供のクレヨン画の夢のような世界がやさしく重なったイメージ。 懐かしく、可愛らしく、少しセンチメンタルで、絵本の中にしまっておきたくなるような一枚。 写真に書かれていない文字は入れない。
GENEL | AIを用いた動画制作 tweet mediaGENEL | AIを用いた動画制作 tweet mediaGENEL | AIを用いた動画制作 tweet mediaGENEL | AIを用いた動画制作 tweet media
日本語
7
130
1.9K
195K
SK
SK@hey2SK·
@ChillaiKalan__ Why the prompt is so long even if you've already used the most advanced model.😂
English
0
0
0
14
K
K@ChillaiKalan__·
GPT IMAGE 2 on ChatGPT Prompt: Create a bold Y2K pop-street editorial collage featuring a young person as the central subject, combining playful graffiti typography, orange poster background, sticker-like graphic elements, and fashion magazine layout energy. MAIN SUBJECT AND FRAMING: Place the young person on the right side of the composition in a close-up to upper-body portrait. The face should be large and dominant, cropped from the chest upward, with the subject slightly leaning forward into the frame. The composition should feel asymmetrical, with graphic typography filling the left side and the subject anchoring the right side. Do not control or describe the hairstyle. OUTFIT DETAIL: Style the subject in a minimal black streetwear outfit. They wear a black turtleneck or high-neck top with a soft fitted silhouette, creating a clean dark shape against the bright orange background. Add a black beret or rounded cap with a small embroidered logo detail on the front, giving the look a stylish retro street-fashion feeling. Add a small colorful cheek sticker or decorative bandage on one cheek, with cute tiny graphic details, making the styling feel playful and idol-like. Keep the outfit mostly black so the blue graffiti typography and orange background stand out strongly. BEHAVIOUR AND BODY LANGUAGE: The subject’s body language should feel calm, closed-off, and slightly moody. The shoulders are relaxed but slightly rounded forward, creating a soft guarded posture. The head tilts subtly downward and to the side, giving the portrait a cool detached attitude. The subject does not need a big hand gesture in the main portrait; instead, the behaviour comes from the face angle, lowered posture, and intense sideways eye movement. The pose should feel like someone caught in a quiet bratty moment, not fully engaging with the viewer. FACIAL EXPRESSION: The expression is pouty, serious, and mildly irritated in a fashionable way. The lips are softly pushed forward into a small pout, brows slightly tense, and the eyes look sideways with a suspicious or unimpressed gaze. The mood should feel bratty, stylish, aloof, and quietly rebellious. GRAPHIC BACKGROUND: Use a saturated orange background with a large white starburst shape behind the subject’s head and shoulders. The starburst should create strong comic-book energy and make the face pop. Add hand-drawn black scribble text across the orange background, like messy diary writing or lyric notes. Include phrases such as: “EENIE MEENY MINEY MO” “4:03” “LOVE DAY” “HIDE YOUR LOVE” “BOYZ” The handwriting should look playful, imperfect, thin, and chaotic, but still visually controlled. MAIN TYPOGRAPHY: Place a large blue graffiti-style word “Your Name” across the left-center area. The letters should be chunky, rounded, irregular, and outlined in thick black, like a playful bubble graffiti tag. The typography should overlap slightly with the subject’s face area without blocking the eyes. Add pink star sparkle stickers around the face and cap area, giving a cute pop-idol accent. BOTTOM FILMSTRIP: At the bottom of the composition, add a horizontal filmstrip with four rectangular photo panels. These panels show smaller full-body and close-up poses of the same young person wearing the same black outfit. Panel pose ideas: 1. Side close-up portrait with head tilted back slightly, cheek sticker visible, eyes looking away. 2. Full-body stance with legs apart, hands near pockets or belt area, posture confident and relaxed. 3. Repeated close-up side portrait for rhythmic collage effect. 4. Repeated full-body stance, slightly different angle or crop. The filmstrip should feel like a fashion contact sheet or idol editorial preview. MOOD AND STYLE: The overall mood should be playful, bratty, nostalgic, and street-fashion inspired. It should feel like a Y2K teen magazine layout mixed with graffiti poster design, idol concept styling, and scrapbook-like pop graphics.
K tweet mediaK tweet mediaK tweet media
English
54
88
537
20.6K
SK
SK@hey2SK·
@venturetwins As a aigc engineer, I am very interested into testing a new image generation model.😘
English
0
0
0
46
Justine Moore
Justine Moore@venturetwins·
I've been testing a new image model dropping soon 👀 It's a breath of fresh air - the visuals are insane, and it makes image generation feel like art again. I have a couple invites, comment or DM if you want to try it!
Justine Moore tweet mediaJustine Moore tweet mediaJustine Moore tweet mediaJustine Moore tweet media
English
118
15
429
53.5K
SK
SK@hey2SK·
Totally Love It. My dream anime.
AIスタジオワンルーム(AIアニメ、動画、漫画)@studio_oneroom

コピペで出来る! 誰でもパスタのプロンプト この動画、簡単に作れます。 キャラクターシートを入れて下記のプロンプトをコピペすれば誰でも簡単にパスタ動画が作れます Seedance2.0用です。 基本的にペペロンチーノですけど、トマトソース風になったりすることもあります。 プロンプト ------ 日本のフルカラーアニメ 高速なカット割り、作画枚数多め セル画調の影。テンポの速い料理シーン。 キャラクターの手と食材の一貫性を維持。 # スタイル 映画品質の料理映像。ウォームホワイトの間接光+スチームによるリムライト。 アナモルフィックレンズフレア、浅い被写界深度、35mmフィルムライクな粒子感。 ティール&オレンジのカラーグレーディング。 # 情景 明るい家庭のキッチン。白いタイル、木製まな板、 銀色のフライパン。昼間の自然光。 # シーン ペペロンチーノを勢いよく作る一連の工程をテンポよく切り取る。 cut1: エクストリームクローズアップ。包丁がニンニクを一気にスライスする。 環境音:リズミカルな包丁音。 Lens switch cut2: マクロ・俯瞰。スライスされたニンニクがまな板に広がる静止ショット。 環境音:静寂。 Lens switch cut3: ローアングル。フライパンにオリーブオイルが注がれ広がる。逆光で輝く。 環境音:サラサラとした液体音。 Lens switch cut4: エクストリームクローズアップ。ニンニクと唐辛子がオイルに投入される瞬間。 環境音:じゅわっという音。 Lens switch cut5: 真横・マクロ。オイルの表面で泡が立ち上がりニンニクが黄金色に変わる。 環境音:じわじわと油が泡立つ音。 Lens switch cut6: 俯瞰。大鍋の中でパスタが湯の渦に泳ぐ。蒸気が上へ流れる。 環境音:ぼこぼこと沸騰する音。 Lens switch cut7: ローアングル・縁スレスレ。パスタがフライパンに投入され蒸気が爆発的に立ち昇る。 環境音:炸裂するような蒸気音。 Lens switch cut8: ミディアムクローズアップ・斜め前から。フライパンを豪快にあおる。 パスタが宙に大きく舞う。 環境音:パスタが空気を切る音、金属音。 Lens switch cut9: バストショット。白い皿にパスタをトングで勢いよく盛り付ける。 環境音:皿とトングが触れる音。 Lens switch cut10: 俯瞰・真上。パセリが指先から舞い落ちて完成した皿の上に着地。 カメラ固定のまま数秒静止。 環境音:完全な静寂。

English
1
0
1
140