Haixun Wang
221 posts

Haixun Wang
@haixunwang
VP Engineering & Head of AI @ EvenUp, ex Instacart, Amazon, Facebook, Google, Microsoft, IBM



True. I think @realDonaldTrump winning makes a big difference in humanity getting to Mars and making life multiplanetary. This might one day save life as we know it.






The ChatGPT desktop app for macOS is now available for all users. Get faster access to ChatGPT to chat about email, screenshots, and anything on your screen with the Option + Space shortcut: openai.com/chatgpt/mac/

“Concepts are the glue that holds our mental world together.”– Murphy (2004) I started to work on conceptualization in 2010 when I joined #MSRA to work on #Probase (haixun.github.io/probase.html) with Haixun Wang @haixunwang . At that time, we leveraged Probase to conceptualize things in the world through contextualization and composition. We started to think what if we can also conceptualize events? I worked with my intern Fangting Xia, however, we met much difficulty. Later when I joined #HKUST, I started my project, #ASER (Activities, States, Events, and their Relations, github.com/HKUST-KnowComp…), with my students Hongming Zhang @hongming110 , Xin Liu, and Haojie Pan. We use IE to find a lot discourse relations for events, and we prove that we can transfer such knowledge to commonsense knowledge such as #TransOMCS (github.com/HKUST-KnowComp…) and #DISCOS (github.com/HKUST-KnowComp…). In ASER, we started to build on our idea of conceptualization. Then we initiated the new project, #AbstractATOMIC (github.com/HKUST-KnowComp…), in which we conceptualize both entities and events in #ATOMIC knowledge base. Now, I am so happy that the paper has been eventually accepted by Artificial Intelligence (doi.org/10.1016/j.arti…). Big thanks to my students Mutian He, Tianqing Fang @TFang229, and Weiqi Wang @MightyWeaver2 . In fact, the review took 2 years, but we have been working on conceptualization without waiting for the outcome. Here is the list of what we have been doing: On the Role of Entity and Event Level Conceptualization in Generalizable Reasoning: A Survey of Tasks, Methods, Applications, and Future Directions AbsInstruct: Eliciting Abstraction Ability from LLMs through Explanation Tuning with Plausibility Estimation (github.com/HKUST-KnowComp…) AbsPyramid: Benchmarking the Abstraction Ability of Language Models with a Unified Entailment Graph (github.com/HKUST-KnowComp…) 🕯️CANDLE: Iterative Conceptualization and Instantiation Distillation from Large Language Models for Commonsense Reasoning (github.com/HKUST-KnowComp…) 🚗CAR: Conceptualization-Augmented Reasoner for Zero-Shot Commonsense Question Answering (github.com/HKUST-KnowComp…) 🐈CAT: A Contextualized Conceptualization and Instantiation Framework for Commonsense Reasoning (github.com/HKUST-KnowComp…) Moving forward, we are now working on something beyond social and physical knowledge reasoning, which we call metaphysical reasoning (github.com/HKUST-KnowComp…). The essential building block of such reasoning is again, conceptualization. I am really excited about the development we have done. It's something like your 10+ year dream coming true. I really appreciate all my collaborators, especially my genius students working on related stuff, trusting me and pushing this direction to be even more interesting!

Here is how to build super intelligence in one straight shot Step 1 - Get a GPU super cluster and train a set of foundation models Step 2 - Have these foundation models curate, clean and generate a bunch of dataset based on logic constraints and heuristics Step 3 - Build the next generation of LLMs based on these datasets. Step 4 - Use the new LLMs to generate datasets that address harder and more complex problems and the steps required to create them. You also need to set up mechanisms to validate these datasets including using humans when it makes sense. Step 5 - Repeat 3 and 4 until you get to super intelligence


SQL Has Been Replaced by English You don't need to learn SQL anymore; you can simply learn English 🤣 Our platform users have been using this feature for the last 18 months.




It’s kinda funny that humans take all this trouble to create complex PDFs with all kinds of figures and tables in it… just so the PDFs can be fed to LLMs to deconstruct these figures and tables back into plain english language 🤷♀️ Soon AI models will create these PDFs and other AI models will deconstruct them - completely eliminate the human in the loop



