DrSinister

4.6K posts

DrSinister banner
DrSinister

DrSinister

@prof_sinister

AI, Art nonArtificial Inteligence at @exploreKeros 🧠 at day @rudzinskimaciej

เข้าร่วม Şubat 2022
437 กำลังติดตาม897 ผู้ติดตาม
ทวีตที่ปักหมุด
DrSinister
DrSinister@prof_sinister·
I'm young on the scene in comparison as I started spring 2020 😉 that's my favourite from spring 2021 I guess modifying how VQGAN works became my post-Covid therapy 🤷 as that was only thing I was able to do - one modification then testing x.com/bit__flip/stat…
Visual Synthesizer - Daniel Joy Grimes@visual_synth

@prof_sinister @nvnot_ Not sure how far back we want to go though. My first text to image was attnGAN back in 2018 or so. it was great for inspiration but was a bit limited. Sleep well, look forward to your story tomorrow :-) "Adversarial Examples" 2018

English
1
1
6
657
DrSinister
DrSinister@prof_sinister·
Banana used to refurbishe old Polish painting of Witold Wojtkiewicz
DrSinister tweet media
Polski
0
0
0
12
DrSinister รีทวีตแล้ว
DrSinister รีทวีตแล้ว
proxima centauri b
proxima centauri b@proximasan·
was she the first case of llm psychosis
proxima centauri b tweet media
English
0
1
7
237
DrSinister
DrSinister@prof_sinister·
@sureailabs BBC? Nature, The Economist It's getting worse every year
English
0
0
0
10
sure, ai
sure, ai@sureailabs·
What are your most credible mainstream new outlets right now? Bonus points if they are Canadian and/or not US-based
English
1
0
0
166
DrSinister
DrSinister@prof_sinister·
@ArtificialAnlys Where's qwen 3 max? That's currently flagship sota Chinese model - it would be good to know where it stands
English
0
0
0
60
Artificial Analysis
Artificial Analysis@ArtificialAnlys·
DeepSeek’s updated V3.1 Terminus ties with gpt-oss-120b (high) as the most intelligent open weights model and offers increased instruction following and long context reasoning capabilities 🧠 Our benchmarking results indicate DeepSeek V3.1 Terminus shows a greater intelligence uplift over DeepSeek V3.1 in reasoning mode compared to non-reasoning mode: ➤ DeepSeek V3.1 Terminus scores 58 in reasoning mode on the Artificial Analysis Intelligence Index, up from V3.1’s score of 54 in reasoning mode. The largest improvements are seen across instruction following (increase of 15 percentage points in IFBench), long context performance (increase of 12 p.p. in AA-LCR) and agentic coding & terminal use (increase of 4 p.p in Terminal-Bench Hard) ➤ In non-reasoning mode, DeepSeek V3.1 Terminus achieves a score of 46, a slight increase over the earlier V3.1 score of 45. 🤖 Other benchmarking takeaways: ➤ Function calling / tool use: Similar to DeepSeek V3.1, V3.1 Terminus does not support function calling when in reasoning mode. This is likely to substantially limit its ability to support agentic workflows with intelligence requirements, including in coding agents. ➤ Token usage: DeepSeek V3.1 Terminus scores higher in reasoning mode than V3.1, and uses more tokens across the evals in the Artificial Analysis Intelligence Index (67M for V3.1 Terminus in reasoning mode vs. 63M for V3.1 in reasoning mode). In non-reasoning mode, V3.1 Terminus uses fewer tokens than V3.1 (11M and 14M respectively). Both V3.1 Terminus and V3.1 use fewer tokens in reasoning mode than DeepSeek’s earlier R1 and R1 0528 reasoning models ➤ Availability: DeepSeek’s first party API now serves the new DeepSeek V3.1 Terminus model on both their chat and reasoning endpoints ➤ Architecture: DeepSeek V3.1 Terminus is architecturally identical to prior V3 and R1 models, with 671B total parameters and 37B active parameters ➤ Providers: There are select third-party providers that are hosting this model such as @DeepInfra (FP4 quantized) and @novita_labs (FP8 quantized)
Artificial Analysis tweet media
English
16
51
482
74.8K
DrSinister
DrSinister@prof_sinister·
@sureailabs @solarapparition Hi hi, check my named account - a bit of magic but we might have solved it - you just need a few h EEG session 🤷
English
0
0
0
17
sure, ai
sure, ai@sureailabs·
@prof_sinister @solarapparition The current archetype of pretty much all assistant AI is to cater to the user's questions rather than asking the user questions to understand what course of action is the best one. While you can easily do this with a good system prompt, no big model is doing that well rn, imo.
English
1
0
3
368
sure, ai
sure, ai@sureailabs·
Story from the customer-facing side of the AI startup that I work at: A company reaching the 12,000 column limit on postgreSQL. it's one big table. Thousands of columns of null values. They hope AI can save them, but the solution is just to be a little more thoughtful.
English
3
0
8
406
Pawel Szczesny
Pawel Szczesny@PawelPSzczesny·
People doubt it or claim the codebase must be in a bad shape, instead of just checking on their own if the tech is capable of really speeding up things. Out of curiosity I tested Hmmer package (in C) developed for 20 years and pretty optimized. Codex found a way to shave another few percentage points (it was told to find one file to optimize so I didn't wait hours) within 30 mins. Measurement was with official benchmarks from the package. This technology is getting better and better.
Pawel Szczesny tweet media
Sauers@Sauers_

I asked Codex to speed up my code by 20x, thinking it would be very difficult. Codex instead sped up my code (already written in Rust) by 250,000x

English
2
0
4
316
DrSinister
DrSinister@prof_sinister·
@sureailabs @solarapparition We all have and probably better use of ai would be to show us the blind spots and what experts see than to automate - don't listen to our questions just say what's the correct choice
English
2
0
3
70
DrSinister
DrSinister@prof_sinister·
I've added an instruction for how to get the "old genAI" vibe with current diffusion models in comments if anybody is interested
nvnot@nvnot_

stuff like this

English
0
0
3
46
DrSinister
DrSinister@prof_sinister·
@nvnot_ run once more as img2img with normal or even high cfg low steps you should get very similar result thats a chack as it could be done in one go with sheduler shenigans
English
0
0
1
17
DrSinister
DrSinister@prof_sinister·
@nvnot_ smalest posible CFG for particular steps count (higher steps -> lower CFG ; there is a match for them and prompt so you need to find it) when you have the img rescale it to make it larger e.g. be 25-50% the largest that the model will be able to use and ...
English
1
0
1
24
DrSinister
DrSinister@prof_sinister·
@nvnot_ do it myself :/ id do algos but implementation is a different skill
English
0
0
1
10
DrSinister
DrSinister@prof_sinister·
@nvnot_ they still can you just need to "manually" change how cfg and noise schedule goes if you know of a person that is willing to read some weird unusual mostly signal analysis code for shedulers then nearly any diffusion model can be remade this way - I just lack the time/skill to...
English
2
0
1
27
DrSinister
DrSinister@prof_sinister·
@torchcompiled 🤔 We should talk about that someday I know a guy working on BCI integration with LLMs based on psychology and philosophy
English
0
0
1
19
Ethan
Ethan@torchcompiled·
New post! I believe we can think of ourselves in two different lenses: an exact point of experience and the history of our patterns of behavior. Though the two are deeply interconnected.
Ethan tweet media
English
3
2
22
2.4K
DrSinister
DrSinister@prof_sinister·
@proximasan You might want to read about cognitive complexity trait - you will like what you will find
English
0
0
1
30
proxima centauri b
proxima centauri b@proximasan·
weird that dichotomous thinking is supposed to be an ASD trait but i often find myself wondering why others are "collapsing spectra into binaries”
English
2
1
5
1.3K