Hanuma

971 posts

Hanuma

Hanuma

@castasstring

Indian software engineer.

Hyderabad, India Katılım Ocak 2018
829 Takip Edilen47 Takipçiler
Sam Bhagwat
Sam Bhagwat@calcsam·
last month we wrote a new agents book: patterns for building ai agents it has everything you need to take your agents from prototype to production, like agent design patterns, the basics of security, etc reply to this tweet with BOOK and we'll dm you so you can get a copy
Sam Bhagwat tweet media
English
4.1K
454
5.1K
587.7K
Hanuma
Hanuma@castasstring·
@IncomeTaxIndia @CAChirag It isn’t working. I’ve been trying to login for a long time, but it doesn’t do anything
English
0
0
0
60
Income Tax India
Income Tax India@IncomeTaxIndia·
Dear @CAChirag, Our checks show that the e-filing portal is working fine. May we request you to share the specific details of the issue you've encountered (along with your mobile number & a screenshot of the error) with us via email at orm@cpc.incometax.gov.in. Our team will get in touch with you.
Income Tax India tweet media
English
140
11
18
47K
CA Chirag Chauhan
CA Chirag Chauhan@CAChirag·
The @IncomeTaxIndia has sent a new SMS stating that ITRs are live on the portal. File today, on e filing portal but their tax portal is down! What to do?
CA Chirag Chauhan tweet media
English
23
142
197
23.7K
Hanuma retweetledi
Tivadar Danka
Tivadar Danka@TivadarDanka·
Problem-solving is at least 50% of every job in tech and science. Mastering problem-solving will make your technical skill level shoot up like a hockey stick. Yet, we are rarely taught how to do so. Here are my favorite techniques that'll loosen even the most complex knots:
Tivadar Danka tweet media
English
8
99
775
65.7K
Hanuma
Hanuma@castasstring·
@realhyderabad86 @SulliRonni61327 NRIs buying using dollars is probably the very reason why 99% of people, working in India and paying taxes here, are not able afford the homes right now. Builders are building to lure NRIs, which is of course their right to do so as well.
English
1
0
4
306
realhyderabad
realhyderabad@realhyderabad86·
@SulliRonni61327 I don't expect corrections of more than 10-15% if at all it happens. The NRI inflows continue unabated. Though not like 2022/23. Many just buy for the sake of it
English
3
0
2
3.9K
realhyderabad
realhyderabad@realhyderabad86·
We Telugus have a deep-rooted, subconscious obsession with real estate. What we witnessed during KCR’s era was just half the picture. The real bubble is yet to form—a time when everything catches fire. At some point, Hyderabad will see peak euphoria—a phase of blind speculation, FOMO-driven buying, and unsustainable price spikes. We won’t know when the euphoria will come. But when it does, it will be legendary.
English
50
20
346
36.1K
neural nets.
neural nets.@cneuralnetwork·
went to the heaven of ML/AI today
neural nets. tweet medianeural nets. tweet media
English
168
157
3.6K
172.7K
Hanuma retweetledi
Legend Prabhas 🇮🇳
Legend Prabhas 🇮🇳@CanadaPrabhasFN·
Neeku unna shakthi lo 0.1% maku ivvavayya chalu🥹 Jai Hanuman🙏🙇🏻❤️
हिन्दी
11
352
4.8K
86.5K
Dr.Sayajirao Gaikwad
Dr.Sayajirao Gaikwad@DietDrsayajirao·
Ragi is diabetes friendly!!! This is misinformation. Ragi has same carbohydrates and is high in calories like rice, wheat and other millets. This chart deserves place in every kitchen to be photo framed. Every family members needs this knowledge of nutritional contents of every foods they’re eating.
Dr.Sayajirao Gaikwad tweet media
English
132
497
2.5K
405.4K
Arundhati Ramanathan
Arundhati Ramanathan@tam_arund·
So @nithyshankar and I adopted a baby girl and are thrilled to be parents to baby N. On a six month maternity break from work— going from making sense of corporate mumbo jumbo to baby babble for now 😄
English
130
14
1.1K
100K
Dr.Krishank
Dr.Krishank@Krishank_BRS·
Request Congress Government not to hold any grudge against this RTC driver but rather understand his suffering.
English
44
424
1.8K
66.3K
Hanuma
Hanuma@castasstring·
We often discuss DevOps, but the 'Ops' world extends beyond just DevOps. Integrating various Ops disciplines, including the emerging trend of LLMOps, with different aspects of product development, is enhancing our ability to create superior software. linkedin.com/posts/hanuma-p…
English
0
0
0
34
Hanuma
Hanuma@castasstring·
@KTRBRS @BRSparty Though I am an admirer of BJP, I doubt whether I’ll see any other dynamic minister as @KTRBRS was…look at the grace with which he is congratulating…
English
0
0
0
9
KTR
KTR@KTRBRS·
Grateful to the people of Telangana for giving @BRSparty two consecutive terms of Government 🙏 Not saddened over the result today, but surely disappointed as it was not in expected lines for us. But we will take this in our stride as a learning and will bounce back Congratulations to Congress party on winning the mandate. Wishing you Good Luck
English
7K
7.5K
59.2K
5.7M
Hanuma
Hanuma@castasstring·
If you are running @playwrightweb tests on AKS using the official docker image, what’s your resource and memory limits? How much memory is needed for it to be running? @debs_obrien any suggestion? Is there a way to print the progress so it can be see? #playwright #playwrightweb
English
0
0
0
20
Hanuma retweetledi
Microsoft Mechanics
Microsoft Mechanics@MSFTMechanics·
No more repetitive questions. Technicians have access to your entire chat history, for a more informed and efficient conversation. Build GPT-automated customer support with Azure Communication Services. youtu.be/N0Cay8md9s4 #AzureServices
YouTube video
YouTube
English
5
9
24
23.3K
Hanuma retweetledi
Bindu Reddy
Bindu Reddy@bindureddy·
LLMs and Their Human-like Reasoning Capabilities - Fact or Fiction? Research into enabling Large Language Models (LLMs) to display human-like reasoning has been a hot topic. However, some others have argued that LLMs aren't capable of reasoning given that are glorified next-word predictors. Here is where we are today - Chain of thought (CoT) reasoning - Prompting a "chain of thought"—a series of intermediate reasoning steps—can significantly improve the performance of LLMs in complex reasoning tasks. The human brain typically decomposes a math problem into intermediate steps and solves each step before giving the final answer: “After Jane gives 2 flowers to her mom she has 10 . . . then after she gives 3 to her dad she will have 7 . . . so the answer is 7.” The basic idea behind CoT is to prompt the model with some examples of these types of problems and the step-by-step reasoning involved in solving them. By prompting in this way, the model can decompose the problem similar to the human brain and solve for it. LLMs when prompted with just eight chain-of-thought examples, achieved state-of-the-art accuracy surpassing even fine-tuned versions of GPT-3 CoT prompting helps in improving airthmetic, commonsense, symbolic and multi-step reasoning capabilities of the LLM That said, the CoT prompting is fairly limited. It depends a lot on the quality of the prompt and the model doesn't have memory and the prompt size is limited by the model's context length. Recently researchers at DeepMind released a paper that outlines the Tree of Thoughts (ToT) framework. It addresses the shortcomings of existing approaches that do not explore different continuations within a thought process or incorporate any type of planning, lookahead, or backtracking to evaluate different options. ToT frames any problem as a search over a tree, where each node represents a partial solution with the input and the sequence of thoughts so far. It allows LMs to explore multiple reasoning paths over thoughts, each thought being a coherent language sequence that serves as an intermediate step toward problem solving. The ToT process involves four key steps: - Decomposing the intermediate process into thought steps. - Generating potential thoughts from each state. - Heuristically evaluating states. - Deciding what search algorithm to use. The ToT framework allows LMs to perform deliberate decision making by considering multiple different reasoning paths and self-evaluating choices to decide the next course of action. It also enables looking ahead or backtracking when necessary to make global choices. The framework is versatile and can handle challenging tasks. It also improves the interpretability of model decisions and the opportunity for human alignment, as the resulting representations are readable, high-level language reasoning instead of implicit, low-level token values. In practice, the ToT framework has significantly enhanced language models' problem-solving abilities on tasks requiring non-trivial planning or search. In the Game of 24, while a model with chain-of-thought prompting only solved 4% of tasks, the ToT method achieved a success rate of 74%. ToT is a prompting framework has well and has the same fundamental limitations as the CoT framework. ToT needs careful thought decomposition and generation, and relies on heuristics for evaluating states and deciding on the search algorithm. However it's important to remember that while LLMs can mimic certain types of reasoning to a certain extent, but it's not the same as human reasoning. They don't have the ability to understand, self-correct, or make judgments based on a deep understanding of the world. They're more like really good actors who can deliver their lines convincingly but don't actually understand the plot of the play. In summary, AI is still extremely early when it comes to reasoning and we may need a another significant breakthrough before LLMs can really measure up to humans.
Bindu Reddy tweet media
English
16
104
509
116.2K